Rev 41: Add some benchmark results for various flush sizes. in http://bzr.arbash-meinel.com/plugins/groupcompress
John Arbash Meinel
john at arbash-meinel.com
Wed Mar 4 22:32:49 GMT 2009
At http://bzr.arbash-meinel.com/plugins/groupcompress
------------------------------------------------------------
revno: 41
revision-id: john at arbash-meinel.com-20090304223243-xrg48jyhczvpkjxc
parent: john at arbash-meinel.com-20090304214211-rg22q09z8queeer0
committer: John Arbash Meinel <john at arbash-meinel.com>
branch nick: groupcompress
timestamp: Wed 2009-03-04 16:32:43 -0600
message:
Add some benchmark results for various flush sizes.
-------------- next part --------------
=== modified file 'groupcompress.py'
--- a/groupcompress.py 2009-03-04 21:42:11 +0000
+++ b/groupcompress.py 2009-03-04 22:32:43 +0000
@@ -780,6 +780,16 @@
keys_to_add.append((key, '%d %d' % (basis_end, end_point),
(record.parents,)))
basis_end = end_point
+ # Interestingly, the sweet spot is 4MB, at 8 and 2 MB the total
+ # size increases... we need a better way of deciding than just
+ # random testing against a given dataset.
+ # 2MB 10.3MB
+ # 3MB 8.4MB
+ # 4MB 8.6MB
+ # 8MB 10.0MB
+ # This effects file content more than other bits, because they
+ # don't get *large* enough to overflow here. (The total compressed
+ # inventory size is only 1.5MB on my test set.)
if basis_end > 1024 * 1024 * 4:
flush()
self._compressor = GroupCompressor(self._delta)
More information about the bazaar-commits
mailing list