Thu, 11 Oct 2007 17:47:04 -0500 mpatch: use int rather than ssize_t
Matt Mackall <mpm@selenic.com> [Thu, 11 Oct 2007 17:47:04 -0500] rev 5458
mpatch: use int rather than ssize_t We might need to be more clever than this to make Py2.5 happy.
Thu, 11 Oct 2007 17:46:06 -0500 osutil: improve portability
Matt Mackall <mpm@selenic.com> [Thu, 11 Oct 2007 17:46:06 -0500] rev 5457
osutil: improve portability - manually inline mode_to_kind - remove unused alloca include - remove fstatat and associated bits It's not obvious that there's an advantage to using fstatat in terms of performance. The race-avoidance properties of fstatat aren't terribly useful to us either. So best to avoid it until we figure out how to use it portably.
Thu, 11 Oct 2007 16:28:30 +0200 fix typo
Benoit Boissinot <benoit.boissinot@ens-lyon.org> [Thu, 11 Oct 2007 16:28:30 +0200] rev 5456
fix typo
Thu, 11 Oct 2007 16:26:14 +0200 import gettext since '_' is used
Benoit Boissinot <benoit.boissinot@ens-lyon.org> [Thu, 11 Oct 2007 16:26:14 +0200] rev 5455
import gettext since '_' is used
Thu, 11 Oct 2007 16:19:12 +0200 explicitely use integer division
Benoit Boissinot <benoit.boissinot@ens-lyon.org> [Thu, 11 Oct 2007 16:19:12 +0200] rev 5454
explicitely use integer division
Thu, 11 Oct 2007 12:16:55 +0200 fix UnboundLocalError, refactor a bit
Benoit Boissinot <benoit.boissinot@ens-lyon.org> [Thu, 11 Oct 2007 12:16:55 +0200] rev 5453
fix UnboundLocalError, refactor a bit bin wasn't defined in all branches (bug introduced in 82b4ff3abbcd
Thu, 11 Oct 2007 00:46:56 -0500 bdiff: tweaks for large files
Matt Mackall <mpm@selenic.com> [Thu, 11 Oct 2007 00:46:56 -0500] rev 5452
bdiff: tweaks for large files - adjust the common line threshold to .1% this speeds up a delta of 7M lines of source from 10m to 40s - adjust the scaling of the hash array down a bit as it was raising the peak memory usage significantly
Thu, 11 Oct 2007 00:46:54 -0500 revlog: break up compression of large deltas
Matt Mackall <mpm@selenic.com> [Thu, 11 Oct 2007 00:46:54 -0500] rev 5451
revlog: break up compression of large deltas Python's zlib apparently makes an internal copy of strings passed to compress(). To avoid this, compress strings 1M at a time, then join them at the end if the result would be smaller than the original. For initial commits of large but compressible files, this cuts peak memory usage nearly in half.
Thu, 11 Oct 2007 00:46:53 -0500 revlog: fix caching of buffer objects
Matt Mackall <mpm@selenic.com> [Thu, 11 Oct 2007 00:46:53 -0500] rev 5450
revlog: fix caching of buffer objects
Thu, 11 Oct 2007 00:46:52 -0500 chunkiter: handle large reads more efficiently
Matt Mackall <mpm@selenic.com> [Thu, 11 Oct 2007 00:46:52 -0500] rev 5449
chunkiter: handle large reads more efficiently - for large reads, don't attempt to read more than necessary - if we've gathered the exact number of bytes needed, avoid a string copy
(0) -3000 -1000 -300 -100 -10 +10 +100 +300 +1000 +3000 +10000 +30000 tip