From fd9bb318a921bb5c4d79eb7d199f52eb67054451 Mon Sep 17 00:00:00 2001 From: root Date: Tue, 13 Nov 2007 11:41:46 +0000 Subject: *** empty log message *** --- Changes | 6 +++--- bench.c | 4 ++-- 2 files changed, 5 insertions(+), 5 deletions(-) diff --git a/Changes b/Changes index ab8f086..d9c34d4 100644 --- a/Changes +++ b/Changes @@ -1,15 +1,15 @@ 2.2 - - switched to GPL v2 or any later version. + - switched to 2-clause bsd with "GPL v2 or any later version" option. - speed up compression by ~10-15% in common cases by some manual unrolling. - - import some compiler tricks from JSON::XS. + - import some compiler tricks from JSON::XS, for further speed-ups. - tune hash functions depending on ULTRA_FAST or VERY_FAST settings. - for typical binary data (e.g. /bin/bash, memory dumps, canterbury corpus etc.), speed is now comparable to fastlz, but with better compression ratio. with ULTRA_FAST, it's typically 3-15% faster than fastlz while still maintaining a similar ratio. (amd64 and core 2 duo, ymmv). thanks a lot for the competition :) - - undo inline assembly, it is no longer helpful. + - undo inline assembly in compressor, it is no longer helpful. - no changes to the decompressor. - use a HLOG of 16 by default now (formerly 15). diff --git a/bench.c b/bench.c index 2c26a8d..4289b8e 100644 --- a/bench.c +++ b/bench.c @@ -42,11 +42,11 @@ int main(void) fclose (f); for (lp = 0; lp < 1000; lp++) { - l = lzf_compress (data, DSIZE, data2, DSIZE*2); s=stamp(); + l = lzf_compress (data, DSIZE, data2, DSIZE*2); //l = fastlz_compress_level (1, data, DSIZE, data2); - j = lzf_decompress (data2, l, data3, DSIZE*2); si[0]=measure(s); + j = lzf_decompress (data2, l, data3, DSIZE*2); printf ("\r%10d (%d) ", si[0], l); if (si[0] < min && si[0] > 0) -- cgit v1.2.3