mm: add !pte_present() check on existing hugetlb_entry callbacks
authorNaoya Horiguchi <n-horiguchi@ah.jp.nec.com>
Fri, 6 Jun 2014 14:00:01 +0000 (10:00 -0400)
committerLinus Torvalds <torvalds@linux-foundation.org>
Fri, 6 Jun 2014 20:21:16 +0000 (13:21 -0700)
The age table walker doesn't check non-present hugetlb entry in common
path, so hugetlb_entry() callbacks must check it.  The reason for this
behavior is that some callers want to handle it in its own way.

[ I think that reason is bogus, btw - it should just do what the regular
  code does, which is to call the "pte_hole()" function for such hugetlb
  entries  - Linus]

However, some callers don't check it now, which causes unpredictable
result, for example when we have a race between migrating hugepage and
reading /proc/pid/numa_maps.  This patch fixes it by adding !pte_present
checks on buggy callbacks.

This bug exists for years and got visible by introducing hugepage
migration.

ChangeLog v2:
- fix if condition (check !pte_present() instead of pte_present())

Reported-by: Sasha Levin <sasha.levin@oracle.com>
Signed-off-by: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>
Cc: Rik van Riel <riel@redhat.com>
Cc: <stable@vger.kernel.org> [3.12+]
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
[ Backported to 3.15.  Signed-off-by: Josh Boyer <jwboyer@fedoraproject.org> ]
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
fs/proc/task_mmu.c
mm/mempolicy.c

index 442177b1119a4528b08210a56f5188b3ef1f2a20..c4b2646b6d7c66f7b1c37d6fbaf033cc5ccaa561 100644 (file)
@@ -1351,7 +1351,7 @@ static int gather_hugetbl_stats(pte_t *pte, unsigned long hmask,
        struct numa_maps *md;
        struct page *page;
 
-       if (pte_none(*pte))
+       if (!pte_present(*pte))
                return 0;
 
        page = pte_page(*pte);
index 78e1472933ea0fce8ee57b94ed3d3d9a3b52b3fc..30cc47f8ffa0571822a58eeb4f1de5ffc10ba63f 100644 (file)
@@ -526,9 +526,13 @@ static void queue_pages_hugetlb_pmd_range(struct vm_area_struct *vma,
        int nid;
        struct page *page;
        spinlock_t *ptl;
+       pte_t entry;
 
        ptl = huge_pte_lock(hstate_vma(vma), vma->vm_mm, (pte_t *)pmd);
-       page = pte_page(huge_ptep_get((pte_t *)pmd));
+       entry = huge_ptep_get((pte_t *)pmd);
+       if (!pte_present(entry))
+               goto unlock;
+       page = pte_page(entry);
        nid = page_to_nid(page);
        if (node_isset(nid, *nodes) == !!(flags & MPOL_MF_INVERT))
                goto unlock;