Re: [PATCH] anobjrmap 1/6 objrmap
From: Martin J. Bligh
Date: Wed Mar 24 2004 - 10:58:48 EST
--Andrea Arcangeli <andrea@xxxxxxx> wrote (on Wednesday, March 24, 2004 07:19:57 +0100):
> On Mon, Mar 22, 2004 at 07:53:02AM -0800, Martin J. Bligh wrote:
>> Just against 2.6.5-rc1 virgin is easiest - that's what I was doing the
>> rest of it against ...
>
> here it is:
>
> http://www.us.kernel.org/pub/linux/kernel/people/andrea/patches/v2.6/2.6.5-rc1/anon-vma-2.6.5-rc2-aa2.gz
> http://www.us.kernel.org/pub/linux/kernel/people/andrea/patches/v2.6/2.6.5-rc1/objrmap-core-2.6.5-rc2-aa2.gz
>
>
Yay, that works ;-) Without the rest of your tree, performance of anon_vma
is almost exactly = anon_mm ... of course all this is under no mem pressure,
I'll have to do some more tests on another machine without infinite ram to
see what happens as we start to reclaim ;-)
Kernbench: (make -j N vmlinux, where N = 2 x num_cpus)
Elapsed System User CPU
2.6.5-rc1 45.75 102.49 577.39 1486.00
2.6.5-rc1-partial 44.84 85.75 576.63 1476.67
2.6.5-rc1-hugh 44.79 83.85 576.71 1474.67
2.6.5-rc1-anon_vma 44.66 83.69 577.14 1479.00
2.6.5-rc1-aa3 44.57 81.57 577.45 1477.67
Kernbench: (make -j N vmlinux, where N = 16 x num_cpus)
Elapsed System User CPU
2.6.5-rc1 46.99 121.95 580.82 1495.33
2.6.5-rc1-partial 45.09 97.16 579.59 1501.00
2.6.5-rc1-hugh 45.00 95.45 579.05 1498.67
2.6.5-rc1-anon_vma 44.90 96.17 579.60 1503.67
2.6.5-rc1-aa3 45.03 93.27 579.84 1494.33
Kernbench: (make -j vmlinux, maximal tasks)
Elapsed System User CPU
2.6.5-rc1 46.96 122.43 580.65 1495.00
2.6.5-rc1-partial 45.18 93.60 579.10 1488.33
2.6.5-rc1-hugh 44.89 91.04 578.49 1490.33
2.6.5-rc1-anon_vma 44.92 91.96 578.86 1493.33
2.6.5-rc1-aa3 44.77 89.29 578.61 1491.33
DISCLAIMER: SPEC(tm) and the benchmark name SDET(tm) are registered
trademarks of the Standard Performance Evaluation Corporation. This
benchmarking was performed for research purposes only, and the run results
are non-compliant and not-comparable with any published results.
Results are shown as percentages of the first set displayed
SDET 1 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 3.0%
2.6.5-rc1-partial 101.4% 1.3%
2.6.5-rc1-hugh 100.0% 2.9%
2.6.5-rc1-anon_vma 101.4% 1.9%
2.6.5-rc1-aa3 104.1% 4.0%
SDET 2 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 1.3%
2.6.5-rc1-partial 107.7% 1.0%
2.6.5-rc1-hugh 108.7% 1.5%
2.6.5-rc1-anon_vma 109.5% 0.7%
2.6.5-rc1-aa3 107.4% 1.3%
SDET 4 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 0.7%
2.6.5-rc1-partial 110.5% 0.6%
2.6.5-rc1-hugh 114.6% 1.3%
2.6.5-rc1-anon_vma 113.3% 0.3%
2.6.5-rc1-aa3 116.1% 1.5%
SDET 8 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 0.9%
2.6.5-rc1-partial 119.4% 0.5%
2.6.5-rc1-hugh 120.2% 1.1%
2.6.5-rc1-anon_vma 119.6% 0.0%
2.6.5-rc1-aa3 124.4% 0.2%
SDET 16 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 0.1%
2.6.5-rc1-partial 118.1% 0.2%
2.6.5-rc1-hugh 119.8% 0.4%
2.6.5-rc1-anon_vma 119.9% 0.8%
2.6.5-rc1-aa3 122.1% 1.1%
SDET 32 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 0.2%
2.6.5-rc1-partial 119.2% 1.0%
2.6.5-rc1-hugh 120.4% 0.4%
2.6.5-rc1-anon_vma 121.8% 0.6%
2.6.5-rc1-aa3 121.1% 0.8%
SDET 64 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 0.3%
2.6.5-rc1-partial 122.1% 0.5%
2.6.5-rc1-hugh 123.5% 0.4%
2.6.5-rc1-anon_vma 123.3% 0.8%
2.6.5-rc1-aa3 123.0% 0.6%
SDET 128 (see disclaimer)
Throughput Std. Dev
2.6.5-rc1 100.0% 0.2%
2.6.5-rc1-partial 123.1% 0.4%
2.6.5-rc1-hugh 124.7% 0.7%
2.6.5-rc1-anon_vma 123.9% 0.3%
2.6.5-rc1-aa3 124.4% 0.1%
For interest's sake, here's the diffprofile for kernbench from
anon_mm to the whole -aa tree ...
3808 25386.7% find_trylock_page
568 2704.8% pgd_alloc
273 74.2% dentry_open
125 11.2% file_move
106 23.1% do_page_cache_readahead
64 0.5% do_anonymous_page
...
-64 -1.0% __copy_to_user_ll
-66 -12.2% .text.lock.file_table
-72 -0.8% __d_lookup
-78 -3.9% path_lookup
-84 -14.9% kmap_atomic
-92 -11.0% pte_alloc_one
-97 -13.7% generic_file_open
-106 -11.2% kmem_cache_free
-121 -13.2% release_pages
-126 -12.6% page_add_rmap
-137 -12.9% clear_page_tables
-212 -7.2% zap_pte_range
-235 -100.0% radix_tree_lookup
-239 -12.5% buffered_rmqueue
-268 -17.8% link_path_walk
-291 -100.0% .text.lock.filemap
-397 -20.8% page_remove_rmap
-398 -100.0% pgd_ctor
-461 -21.6% do_no_page
-669 -1.4% default_idle
-3508 -2.5% total
-3719 -99.4% find_get_page
zap_pte_range and page_remove_rmap and do_no_page are cheaper ... are we
setting up and tearing down pages less frequently somehow? Would be
curious to know which patch that is ...
M.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/