[PATCH] mm: rmap with inner ptlock
rmap's page_check_address descend without page_table_lock. First just pte_offset_map in case there's no pte present worth locking for, then take page_table_lock for the full check, and pass ptl back to caller in the same style as pte_offset_map_lock. __xip_unmap, page_referenced_one and try_to_unmap_one use pte_unmap_unlock. try_to_unmap_cluster also. page_check_address reformatted to avoid progressive indentation. No use is made of its one error code, return NULL when it fails. Signed-off-by: Hugh Dickins <hugh@veritas.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
67b02f119d
commit
c0718806cf
3 changed files with 60 additions and 65 deletions
|
@ -95,8 +95,8 @@ int try_to_unmap(struct page *);
|
||||||
/*
|
/*
|
||||||
* Called from mm/filemap_xip.c to unmap empty zero page
|
* Called from mm/filemap_xip.c to unmap empty zero page
|
||||||
*/
|
*/
|
||||||
pte_t *page_check_address(struct page *, struct mm_struct *, unsigned long);
|
pte_t *page_check_address(struct page *, struct mm_struct *,
|
||||||
|
unsigned long, spinlock_t **);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Used by swapoff to help locate where page is expected in vma.
|
* Used by swapoff to help locate where page is expected in vma.
|
||||||
|
|
|
@ -174,6 +174,7 @@ __xip_unmap (struct address_space * mapping,
|
||||||
unsigned long address;
|
unsigned long address;
|
||||||
pte_t *pte;
|
pte_t *pte;
|
||||||
pte_t pteval;
|
pte_t pteval;
|
||||||
|
spinlock_t *ptl;
|
||||||
struct page *page;
|
struct page *page;
|
||||||
|
|
||||||
spin_lock(&mapping->i_mmap_lock);
|
spin_lock(&mapping->i_mmap_lock);
|
||||||
|
@ -183,20 +184,15 @@ __xip_unmap (struct address_space * mapping,
|
||||||
((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
|
((pgoff - vma->vm_pgoff) << PAGE_SHIFT);
|
||||||
BUG_ON(address < vma->vm_start || address >= vma->vm_end);
|
BUG_ON(address < vma->vm_start || address >= vma->vm_end);
|
||||||
page = ZERO_PAGE(address);
|
page = ZERO_PAGE(address);
|
||||||
/*
|
pte = page_check_address(page, mm, address, &ptl);
|
||||||
* We need the page_table_lock to protect us from page faults,
|
if (pte) {
|
||||||
* munmap, fork, etc...
|
|
||||||
*/
|
|
||||||
pte = page_check_address(page, mm, address);
|
|
||||||
if (!IS_ERR(pte)) {
|
|
||||||
/* Nuke the page table entry. */
|
/* Nuke the page table entry. */
|
||||||
flush_cache_page(vma, address, pte_pfn(*pte));
|
flush_cache_page(vma, address, pte_pfn(*pte));
|
||||||
pteval = ptep_clear_flush(vma, address, pte);
|
pteval = ptep_clear_flush(vma, address, pte);
|
||||||
page_remove_rmap(page);
|
page_remove_rmap(page);
|
||||||
dec_mm_counter(mm, file_rss);
|
dec_mm_counter(mm, file_rss);
|
||||||
BUG_ON(pte_dirty(pteval));
|
BUG_ON(pte_dirty(pteval));
|
||||||
pte_unmap(pte);
|
pte_unmap_unlock(pte, ptl);
|
||||||
spin_unlock(&mm->page_table_lock);
|
|
||||||
page_cache_release(page);
|
page_cache_release(page);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
85
mm/rmap.c
85
mm/rmap.c
|
@ -247,34 +247,41 @@ unsigned long page_address_in_vma(struct page *page, struct vm_area_struct *vma)
|
||||||
* On success returns with mapped pte and locked mm->page_table_lock.
|
* On success returns with mapped pte and locked mm->page_table_lock.
|
||||||
*/
|
*/
|
||||||
pte_t *page_check_address(struct page *page, struct mm_struct *mm,
|
pte_t *page_check_address(struct page *page, struct mm_struct *mm,
|
||||||
unsigned long address)
|
unsigned long address, spinlock_t **ptlp)
|
||||||
{
|
{
|
||||||
pgd_t *pgd;
|
pgd_t *pgd;
|
||||||
pud_t *pud;
|
pud_t *pud;
|
||||||
pmd_t *pmd;
|
pmd_t *pmd;
|
||||||
pte_t *pte;
|
pte_t *pte;
|
||||||
|
spinlock_t *ptl;
|
||||||
|
|
||||||
/*
|
|
||||||
* We need the page_table_lock to protect us from page faults,
|
|
||||||
* munmap, fork, etc...
|
|
||||||
*/
|
|
||||||
spin_lock(&mm->page_table_lock);
|
|
||||||
pgd = pgd_offset(mm, address);
|
pgd = pgd_offset(mm, address);
|
||||||
if (likely(pgd_present(*pgd))) {
|
if (!pgd_present(*pgd))
|
||||||
|
return NULL;
|
||||||
|
|
||||||
pud = pud_offset(pgd, address);
|
pud = pud_offset(pgd, address);
|
||||||
if (likely(pud_present(*pud))) {
|
if (!pud_present(*pud))
|
||||||
|
return NULL;
|
||||||
|
|
||||||
pmd = pmd_offset(pud, address);
|
pmd = pmd_offset(pud, address);
|
||||||
if (likely(pmd_present(*pmd))) {
|
if (!pmd_present(*pmd))
|
||||||
|
return NULL;
|
||||||
|
|
||||||
pte = pte_offset_map(pmd, address);
|
pte = pte_offset_map(pmd, address);
|
||||||
if (likely(pte_present(*pte) &&
|
/* Make a quick check before getting the lock */
|
||||||
page_to_pfn(page) == pte_pfn(*pte)))
|
if (!pte_present(*pte)) {
|
||||||
return pte;
|
|
||||||
pte_unmap(pte);
|
pte_unmap(pte);
|
||||||
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
ptl = &mm->page_table_lock;
|
||||||
|
spin_lock(ptl);
|
||||||
|
if (pte_present(*pte) && page_to_pfn(page) == pte_pfn(*pte)) {
|
||||||
|
*ptlp = ptl;
|
||||||
|
return pte;
|
||||||
}
|
}
|
||||||
}
|
pte_unmap_unlock(pte, ptl);
|
||||||
spin_unlock(&mm->page_table_lock);
|
return NULL;
|
||||||
return ERR_PTR(-ENOENT);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -287,28 +294,28 @@ static int page_referenced_one(struct page *page,
|
||||||
struct mm_struct *mm = vma->vm_mm;
|
struct mm_struct *mm = vma->vm_mm;
|
||||||
unsigned long address;
|
unsigned long address;
|
||||||
pte_t *pte;
|
pte_t *pte;
|
||||||
|
spinlock_t *ptl;
|
||||||
int referenced = 0;
|
int referenced = 0;
|
||||||
|
|
||||||
address = vma_address(page, vma);
|
address = vma_address(page, vma);
|
||||||
if (address == -EFAULT)
|
if (address == -EFAULT)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
pte = page_check_address(page, mm, address);
|
pte = page_check_address(page, mm, address, &ptl);
|
||||||
if (!IS_ERR(pte)) {
|
if (!pte)
|
||||||
|
goto out;
|
||||||
|
|
||||||
if (ptep_clear_flush_young(vma, address, pte))
|
if (ptep_clear_flush_young(vma, address, pte))
|
||||||
referenced++;
|
referenced++;
|
||||||
|
|
||||||
/* Pretend the page is referenced if the task has the
|
/* Pretend the page is referenced if the task has the
|
||||||
swap token and is in the middle of a page fault. */
|
swap token and is in the middle of a page fault. */
|
||||||
if (mm != current->mm && !ignore_token &&
|
if (mm != current->mm && !ignore_token && has_swap_token(mm) &&
|
||||||
has_swap_token(mm) &&
|
|
||||||
rwsem_is_locked(&mm->mmap_sem))
|
rwsem_is_locked(&mm->mmap_sem))
|
||||||
referenced++;
|
referenced++;
|
||||||
|
|
||||||
(*mapcount)--;
|
(*mapcount)--;
|
||||||
pte_unmap(pte);
|
pte_unmap_unlock(pte, ptl);
|
||||||
spin_unlock(&mm->page_table_lock);
|
|
||||||
}
|
|
||||||
out:
|
out:
|
||||||
return referenced;
|
return referenced;
|
||||||
}
|
}
|
||||||
|
@ -507,14 +514,15 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma)
|
||||||
unsigned long address;
|
unsigned long address;
|
||||||
pte_t *pte;
|
pte_t *pte;
|
||||||
pte_t pteval;
|
pte_t pteval;
|
||||||
|
spinlock_t *ptl;
|
||||||
int ret = SWAP_AGAIN;
|
int ret = SWAP_AGAIN;
|
||||||
|
|
||||||
address = vma_address(page, vma);
|
address = vma_address(page, vma);
|
||||||
if (address == -EFAULT)
|
if (address == -EFAULT)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
pte = page_check_address(page, mm, address);
|
pte = page_check_address(page, mm, address, &ptl);
|
||||||
if (IS_ERR(pte))
|
if (!pte)
|
||||||
goto out;
|
goto out;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -564,8 +572,7 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma)
|
||||||
page_cache_release(page);
|
page_cache_release(page);
|
||||||
|
|
||||||
out_unmap:
|
out_unmap:
|
||||||
pte_unmap(pte);
|
pte_unmap_unlock(pte, ptl);
|
||||||
spin_unlock(&mm->page_table_lock);
|
|
||||||
out:
|
out:
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
@ -599,19 +606,14 @@ static void try_to_unmap_cluster(unsigned long cursor,
|
||||||
pgd_t *pgd;
|
pgd_t *pgd;
|
||||||
pud_t *pud;
|
pud_t *pud;
|
||||||
pmd_t *pmd;
|
pmd_t *pmd;
|
||||||
pte_t *pte, *original_pte;
|
pte_t *pte;
|
||||||
pte_t pteval;
|
pte_t pteval;
|
||||||
|
spinlock_t *ptl;
|
||||||
struct page *page;
|
struct page *page;
|
||||||
unsigned long address;
|
unsigned long address;
|
||||||
unsigned long end;
|
unsigned long end;
|
||||||
unsigned long pfn;
|
unsigned long pfn;
|
||||||
|
|
||||||
/*
|
|
||||||
* We need the page_table_lock to protect us from page faults,
|
|
||||||
* munmap, fork, etc...
|
|
||||||
*/
|
|
||||||
spin_lock(&mm->page_table_lock);
|
|
||||||
|
|
||||||
address = (vma->vm_start + cursor) & CLUSTER_MASK;
|
address = (vma->vm_start + cursor) & CLUSTER_MASK;
|
||||||
end = address + CLUSTER_SIZE;
|
end = address + CLUSTER_SIZE;
|
||||||
if (address < vma->vm_start)
|
if (address < vma->vm_start)
|
||||||
|
@ -621,22 +623,22 @@ static void try_to_unmap_cluster(unsigned long cursor,
|
||||||
|
|
||||||
pgd = pgd_offset(mm, address);
|
pgd = pgd_offset(mm, address);
|
||||||
if (!pgd_present(*pgd))
|
if (!pgd_present(*pgd))
|
||||||
goto out_unlock;
|
return;
|
||||||
|
|
||||||
pud = pud_offset(pgd, address);
|
pud = pud_offset(pgd, address);
|
||||||
if (!pud_present(*pud))
|
if (!pud_present(*pud))
|
||||||
goto out_unlock;
|
return;
|
||||||
|
|
||||||
pmd = pmd_offset(pud, address);
|
pmd = pmd_offset(pud, address);
|
||||||
if (!pmd_present(*pmd))
|
if (!pmd_present(*pmd))
|
||||||
goto out_unlock;
|
return;
|
||||||
|
|
||||||
|
pte = pte_offset_map_lock(mm, pmd, address, &ptl);
|
||||||
|
|
||||||
/* Update high watermark before we lower rss */
|
/* Update high watermark before we lower rss */
|
||||||
update_hiwater_rss(mm);
|
update_hiwater_rss(mm);
|
||||||
|
|
||||||
for (original_pte = pte = pte_offset_map(pmd, address);
|
for (; address < end; pte++, address += PAGE_SIZE) {
|
||||||
address < end; pte++, address += PAGE_SIZE) {
|
|
||||||
|
|
||||||
if (!pte_present(*pte))
|
if (!pte_present(*pte))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
@ -669,10 +671,7 @@ static void try_to_unmap_cluster(unsigned long cursor,
|
||||||
dec_mm_counter(mm, file_rss);
|
dec_mm_counter(mm, file_rss);
|
||||||
(*mapcount)--;
|
(*mapcount)--;
|
||||||
}
|
}
|
||||||
|
pte_unmap_unlock(pte - 1, ptl);
|
||||||
pte_unmap(original_pte);
|
|
||||||
out_unlock:
|
|
||||||
spin_unlock(&mm->page_table_lock);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static int try_to_unmap_anon(struct page *page)
|
static int try_to_unmap_anon(struct page *page)
|
||||||
|
|
Loading…
Reference in a new issue