Re: Solaris ZFS on Linux [Was: Re: the " 'official' point of view" expressed by regarding reiser4 inclusion]

From: David Lang
Date: Mon Jul 31 2006 - 19:27:48 EST

On Mon, 31 Jul 2006, Nate Diller wrote:

On 7/31/06, Matthias Andree <matthias.andree@xxxxxx> wrote:
Adrian Ulrich wrote:

> See also:
> A quick'n'dirty ZFS-vs-UFS-vs-Reiser3-vs-Reiser4-vs-Ext3 'benchmark'

Whatever Postmark does, this looks pretty besides the point.

why's that? postmark is one of the standard benchmarks...

Are these actual transactions with the "D"urability guarantee?
3000/s doesn't look too much like you're doing synchronous I/O (else
figures around 70/s perhaps 100/s would be more adequate), and cache
exercise is rather irrelevant for databases that manage real (=valuable)

204.62 megabytes read (8.53 megabytes per second)
271.49 megabytes written (11.31 megabytes per second)

looks pretty I/O bound to me, 11.31 MB/s isn't exactly your latest DDR
RAM bandwidth. as far as the synchronous I/O question, Reiser4 in
this case acts more like a log-based FS. That allows it to "overlap"
synchronous operations that are being submitted by multiple threads.

what you are missing is that apps that need to do lots of syncing (databases, mail servers) need to wait for the data to hit non-volitile media before the write is complete. this limits such apps to ~1 write per revolution of the platters (yes it's possible for a limited time to have multiple writes to different things happen to be on the same track, but the counter is the extra seek time needed between tracks)

so any benchmark that shows more transactions then the media has revolutions is highly suspect (now if you have battery-backed cache, or the equivalent you can blow past these limits)

on consumer (7200 rpm) drives this limit is 120/sec, on high-end drives (15Krpm scsi's this is 250/sec, and on the 10k rpm drives in the middle it's about 166/sec.

David Lang
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at
Please read the FAQ at