summaryrefslogtreecommitdiff
path: root/mm
diff options
context:
space:
mode:
authorWaiman Long <Waiman.Long@hp.com>2014-08-06 16:05:36 -0700
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>2014-10-09 12:21:27 -0700
commit1da286ebc5a1d23d0b4b88ba0d64fc141ac4c37d (patch)
tree54b3fc06f70abe1139241a4483be8d21456a18d5 /mm
parentde1fc405fbc586005607e51599da5997463fbefc (diff)
downloadlinux-stable-1da286ebc5a1d23d0b4b88ba0d64fc141ac4c37d.tar.gz
linux-stable-1da286ebc5a1d23d0b4b88ba0d64fc141ac4c37d.tar.bz2
linux-stable-1da286ebc5a1d23d0b4b88ba0d64fc141ac4c37d.zip
mm, thp: move invariant bug check out of loop in __split_huge_page_map
commit f8303c2582b889351e261ff18c4d8eb197a77db2 upstream. In __split_huge_page_map(), the check for page_mapcount(page) is invariant within the for loop. Because of the fact that the macro is implemented using atomic_read(), the redundant check cannot be optimized away by the compiler leading to unnecessary read to the page structure. This patch moves the invariant bug check out of the loop so that it will be done only once. On a 3.16-rc1 based kernel, the execution time of a microbenchmark that broke up 1000 transparent huge pages using munmap() had an execution time of 38,245us and 38,548us with and without the patch respectively. The performance gain is about 1%. Signed-off-by: Waiman Long <Waiman.Long@hp.com> Acked-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Mel Gorman <mgorman@suse.de> Cc: Rik van Riel <riel@redhat.com> Cc: Scott J Norton <scott.norton@hp.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/huge_memory.c4
1 files changed, 2 insertions, 2 deletions
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index 1c42d0c36d0b..4a0c99870774 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -1819,6 +1819,8 @@ static int __split_huge_page_map(struct page *page,
if (pmd) {
pgtable = pgtable_trans_huge_withdraw(mm, pmd);
pmd_populate(mm, &_pmd, pgtable);
+ if (pmd_write(*pmd))
+ BUG_ON(page_mapcount(page) != 1);
haddr = address;
for (i = 0; i < HPAGE_PMD_NR; i++, haddr += PAGE_SIZE) {
@@ -1828,8 +1830,6 @@ static int __split_huge_page_map(struct page *page,
entry = maybe_mkwrite(pte_mkdirty(entry), vma);
if (!pmd_write(*pmd))
entry = pte_wrprotect(entry);
- else
- BUG_ON(page_mapcount(page) != 1);
if (!pmd_young(*pmd))
entry = pte_mkold(entry);
if (pmd_numa(*pmd))