OK, so test2-mm1 fixes the panic I was seeing in test1-mm1.
Only noticeable thing is that -mm tree is consistently a little slower
at kernbench
Kernbench: (make -j N vmlinux, where N = 2 x num_cpus)
Elapsed System User CPU
2.5.74 45.17 97.88 568.43 1474.75
2.5.74-mm1 45.84 109.66 568.05 1477.50
2.6.0-test1 45.25 98.63 568.45 1473.50
2.6.0-test2-mm1 45.38 101.47 569.16 1476.25
2.6.0-test2-mjb1 43.31 75.98 564.33 1478.00
Kernbench: (make -j N vmlinux, where N = 16 x num_cpus)
Elapsed System User CPU
2.5.74 45.74 114.56 571.62 1500.00
2.5.74-mm1 46.59 133.65 570.90 1511.50
2.6.0-test1 45.68 114.68 571.70 1503.00
2.6.0-test2-mm1 46.66 119.82 579.32 1497.25
2.6.0-test2-mjb1 44.03 87.85 569.97 1493.75
Kernbench: (make -j vmlinux, maximal tasks)
Elapsed System User CPU
2.5.74 46.11 115.86 571.77 1491.50
2.5.74-mm1 47.13 139.07 571.52 1509.25
2.6.0-test1 46.09 115.76 571.74 1491.25
2.6.0-test2-mm1 46.95 121.18 582.00 1497.50
2.6.0-test2-mjb1 44.08 85.54 570.57 1487.25
DISCLAIMER: SPEC(tm) and the benchmark name SDET(tm) are registered
trademarks of the Standard Performance Evaluation Corporation. This
benchmarking was performed for research purposes only, and the run results
are non-compliant and not-comparable with any published results.
Results are shown as percentages of the first set displayed
SDET 1 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 3.7%
2.5.74-mm1 88.5% 10.9%
2.6.0-test1 103.0% 2.0%
2.6.0-test2-mm1 99.7% 3.1%
2.6.0-test2-mjb1 107.2% 3.6%
SDET 2 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 53.7%
2.5.74-mm1 133.9% 1.4%
2.6.0-test1 136.4% 1.9%
2.6.0-test2-mm1 132.1% 4.2%
2.6.0-test2-mjb1 156.6% 1.1%
SDET 4 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 3.9%
2.5.74-mm1 92.5% 2.5%
2.6.0-test1 96.7% 5.7%
2.6.0-test2-mm1 70.6% 49.1%
2.6.0-test2-mjb1 134.2% 2.1%
SDET 8 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 45.9%
2.5.74-mm1 123.5% 0.6%
2.6.0-test1 86.1% 70.7%
2.6.0-test2-mm1 127.8% 0.4%
2.6.0-test2-mjb1 158.6% 0.7%
SDET 16 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 0.3%
2.5.74-mm1 92.8% 0.8%
2.6.0-test1 99.3% 0.6%
2.6.0-test2-mm1 97.9% 0.5%
2.6.0-test2-mjb1 120.8% 0.6%
SDET 32 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 0.1%
2.5.74-mm1 94.4% 0.4%
2.6.0-test1 100.4% 0.2%
2.6.0-test2-mm1 97.9% 0.2%
2.6.0-test2-mjb1 123.2% 0.5%
SDET 64 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 0.4%
2.5.74-mm1 95.6% 0.3%
2.6.0-test1 101.1% 0.3%
2.6.0-test2-mm1 100.3% 0.5%
2.6.0-test2-mjb1 127.1% 0.2%
SDET 128 (see disclaimer)
Throughput Std. Dev
2.5.74 100.0% 0.1%
2.5.74-mm1 97.6% 0.2%
2.6.0-test1 100.6% 0.6%
2.6.0-test2-mm1 101.8% 0.0%
2.6.0-test2-mjb1 127.9% 0.3%
diffprofile for kernbench (from test1 to test2-mm1, so not really
fair, but might help):
4383 2.6% total
1600 6.8% page_remove_rmap
934 61.6% do_no_page
470 13.9% __copy_from_user_ll
469 12.8% find_get_page
373 0.0% pgd_ctor
368 4.7% __d_lookup
349 6.6% __copy_to_user_ll
278 15.1% atomic_dec_and_lock
273 154.2% may_open
240 15.6% kmem_cache_free
182 30.2% __wake_up
163 11.2% schedule
152 10.4% free_hot_cold_page
148 21.2% pte_alloc_one
123 6.5% path_lookup
100 9.8% clear_page_tables
77 12.5% copy_process
76 4.2% buffered_rmqueue
70 19.0% .text.lock.file_table
70 5.8% release_pages
66 825.0% free_percpu
55 21.0% vfs_read
54 300.0% cache_grow
50 9.5% kmap_atomic
....
-57 -38.0% __generic_file_aio_read
-62 -100.0% free_pages_bulk
-255 -77.5% dentry_open
-316 -2.2% do_anonymous_page
-415 -77.3% do_page_cache_readahead
-562 -96.1% pgd_alloc
-683 -68.9% filemap_nopage
-1005 -2.0% default_idle
Someone messing with the pgd alloc stuff, perhaps?
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
This archive was generated by hypermail 2b29 : Thu Jul 31 2003 - 22:00:45 EST