#include #include #include #include #include #include "ldm.h" #include "ldm_hashtable.h" #define LDM_HASHTABLESIZE (1 << (LDM_MEMORY_USAGE)) #define LDM_HASHTABLESIZE_U32 ((LDM_HASHTABLESIZE) >> 2) #define LDM_HASHTABLESIZE_U64 ((LDM_HASHTABLESIZE) >> 3) // Insert every (HASH_ONLY_EVERY + 1) into the hash table. #define HASH_ONLY_EVERY_LOG (LDM_WINDOW_SIZE_LOG-((LDM_MEMORY_USAGE)-(LDM_HASH_ENTRY_SIZE_LOG))) #define HASH_ONLY_EVERY ((1 << HASH_ONLY_EVERY_LOG) - 1) #define ML_BITS 4 #define ML_MASK ((1U<>= 1) { ret++; } return ret; } // TODO: Maybe we would eventually prefer to have linear rather than // exponential buckets. /** void HASH_outputTableOffsetHistogram(const LDM_CCtx *cctx) { U32 i = 0; int buckets[32] = { 0 }; printf("\n"); printf("Hash table histogram\n"); for (; i < HASH_getSize(cctx->hashTable); i++) { int offset = (cctx->ip - cctx->ibase) - HASH_getEntryFromHash(cctx->hashTable, i)->offset; buckets[intLog2(offset)]++; } i = 0; for (; i < 32; i++) { printf("2^%*d: %10u %6.3f%%\n", 2, i, buckets[i], 100.0 * (double) buckets[i] / (double) HASH_getSize(cctx->hashTable)); } printf("\n"); } */ void LDM_printCompressStats(const LDM_compressStats *stats) { int i = 0; printf("=====================\n"); printf("Compression statistics\n"); printf("Window size, hash table size (bytes): 2^%u, 2^%u\n", stats->windowSizeLog, stats->hashTableSizeLog); printf("num matches, total match length, %% matched: %u, %llu, %.3f\n", stats->numMatches, stats->totalMatchLength, 100.0 * (double)stats->totalMatchLength / (double)(stats->totalMatchLength + stats->totalLiteralLength)); printf("avg match length: %.1f\n", ((double)stats->totalMatchLength) / (double)stats->numMatches); printf("avg literal length, total literalLength: %.1f, %llu\n", ((double)stats->totalLiteralLength) / (double)stats->numMatches, stats->totalLiteralLength); printf("avg offset length: %.1f\n", ((double)stats->totalOffset) / (double)stats->numMatches); printf("min offset, max offset: %u, %u\n", stats->minOffset, stats->maxOffset); printf("\n"); printf("offset histogram: offset, num matches, %% of matches\n"); for (; i <= intLog2(stats->maxOffset); i++) { printf("2^%*d: %10u %6.3f%%\n", 2, i, stats->offsetHistogram[i], 100.0 * (double) stats->offsetHistogram[i] / (double) stats->numMatches); } printf("\n"); printf("=====================\n"); } int LDM_isValidMatch(const BYTE *pIn, const BYTE *pMatch) { U32 lengthLeft = LDM_MIN_MATCH_LENGTH; const BYTE *curIn = pIn; const BYTE *curMatch = pMatch; if (pIn - pMatch > LDM_WINDOW_SIZE) { return 0; } for (; lengthLeft >= 4; lengthLeft -= 4) { if (MEM_read32(curIn) != MEM_read32(curMatch)) { return 0; } curIn += 4; curMatch += 4; } return 1; } /** * Convert a sum computed from getChecksum to a hash value in the range * of the hash table. */ static hash_t checksumToHash(U32 sum) { return HASH_hashU32(sum); } /** * Computes a 32-bit checksum based on rsync's checksum. * * a(k,l) = \sum_{i = k}^l x_i (mod M) * b(k,l) = \sum_{i = k}^l ((l - i + 1) * x_i) (mod M) * checksum(k,l) = a(k,l) + 2^{16} * b(k,l) */ static checksum_t getChecksum(const BYTE *buf, U32 len) { U32 i; checksum_t s1, s2; s1 = s2 = 0; for (i = 0; i < (len - 4); i += 4) { s2 += (4 * (s1 + buf[i])) + (3 * buf[i + 1]) + (2 * buf[i + 2]) + (buf[i + 3]) + (10 * CHECKSUM_CHAR_OFFSET); s1 += buf[i] + buf[i + 1] + buf[i + 2] + buf[i + 3] + + (4 * CHECKSUM_CHAR_OFFSET); } for(; i < len; i++) { s1 += buf[i] + CHECKSUM_CHAR_OFFSET; s2 += s1; } return (s1 & 0xffff) + (s2 << 16); } /** * Update a checksum computed from getChecksum(data, len). * * The checksum can be updated along its ends as follows: * a(k+1, l+1) = (a(k,l) - x_k + x_{l+1}) (mod M) * b(k+1, l+1) = (b(k,l) - (l-k+1)*x_k + (a(k+1,l+1)) (mod M) * * Thus toRemove should correspond to data[0]. */ static checksum_t updateChecksum(checksum_t sum, U32 len, BYTE toRemove, BYTE toAdd) { U32 s1 = (sum & 0xffff) - toRemove + toAdd; U32 s2 = (sum >> 16) - ((toRemove + CHECKSUM_CHAR_OFFSET) * len) + s1; return (s1 & 0xffff) + (s2 << 16); } /** * Update cctx->nextSum, cctx->nextHash, and cctx->nextPosHashed * based on cctx->lastSum and cctx->lastPosHashed. * * This uses a rolling hash and requires that the last position hashed * corresponds to cctx->nextIp - step. */ static void setNextHash(LDM_CCtx *cctx) { #ifdef RUN_CHECKS U32 check; if ((cctx->nextIp - cctx->ibase != 1) && (cctx->nextIp - cctx->DEBUG_setNextHash != 1)) { printf("CHECK debug fail: %zu %zu\n", cctx->nextIp - cctx->ibase, cctx->DEBUG_setNextHash - cctx->ibase); } cctx->DEBUG_setNextHash = cctx->nextIp; #endif // cctx->nextSum = getChecksum((const char *)cctx->nextIp, LDM_HASH_LENGTH); cctx->nextSum = updateChecksum( cctx->lastSum, LDM_HASH_LENGTH, cctx->lastPosHashed[0], cctx->lastPosHashed[LDM_HASH_LENGTH]); cctx->nextPosHashed = cctx->nextIp; cctx->nextHash = checksumToHash(cctx->nextSum); #if LDM_LAG if (cctx->ip - cctx->ibase > LDM_LAG) { cctx->lagSum = updateChecksum( cctx->lagSum, LDM_HASH_LENGTH, cctx->lagIp[0], cctx->lagIp[LDM_HASH_LENGTH]); cctx->lagIp++; cctx->lagHash = checksumToHash(cctx->lagSum); } #endif #ifdef RUN_CHECKS check = getChecksum(cctx->nextIp, LDM_HASH_LENGTH); if (check != cctx->nextSum) { printf("CHECK: setNextHash failed %u %u\n", check, cctx->nextSum); } if ((cctx->nextIp - cctx->lastPosHashed) != 1) { printf("setNextHash: nextIp != lastPosHashed + 1. %zu %zu %zu\n", cctx->nextIp - cctx->ibase, cctx->lastPosHashed - cctx->ibase, cctx->ip - cctx->ibase); } #endif } static void putHashOfCurrentPositionFromHash( LDM_CCtx *cctx, hash_t hash, U32 checksum) { // Hash only every HASH_ONLY_EVERY times, based on cctx->ip. // Note: this works only when cctx->step is 1. if (((cctx->ip - cctx->ibase) & HASH_ONLY_EVERY) == HASH_ONLY_EVERY) { #if LDM_LAG // TODO: off by 1, but whatever if (cctx->lagIp - cctx->ibase > 0) { const LDM_hashEntry entry = { cctx->lagIp - cctx->ibase, cctx->lagSum }; HASH_insert(cctx->hashTable, cctx->lagHash, entry); } else { const LDM_hashEntry entry = { cctx->ip - cctx->ibase, checksum }; HASH_insert(cctx->hashTable, hash, entry); } #else const LDM_hashEntry entry = { cctx->ip - cctx->ibase, checksum }; HASH_insert(cctx->hashTable, hash, entry); #endif } cctx->lastPosHashed = cctx->ip; cctx->lastHash = hash; cctx->lastSum = checksum; } /** * Copy over the cctx->lastHash, cctx->lastSum, and cctx->lastPosHashed * fields from the "next" fields. * * This requires that cctx->ip == cctx->nextPosHashed. */ static void LDM_updateLastHashFromNextHash(LDM_CCtx *cctx) { #ifdef RUN_CHECKS if (cctx->ip != cctx->nextPosHashed) { printf("CHECK failed: updateLastHashFromNextHash %zu\n", cctx->ip - cctx->ibase); } #endif putHashOfCurrentPositionFromHash(cctx, cctx->nextHash, cctx->nextSum); } /** * Insert hash of the current position into the hash table. */ static void LDM_putHashOfCurrentPosition(LDM_CCtx *cctx) { checksum_t sum = getChecksum(cctx->ip, LDM_HASH_LENGTH); hash_t hash = checksumToHash(sum); #ifdef RUN_CHECKS if (cctx->nextPosHashed != cctx->ip && (cctx->ip != cctx->ibase)) { printf("CHECK failed: putHashOfCurrentPosition %zu\n", cctx->ip - cctx->ibase); } #endif putHashOfCurrentPositionFromHash(cctx, hash, sum); } U32 LDM_countMatchLength(const BYTE *pIn, const BYTE *pMatch, const BYTE *pInLimit) { const BYTE * const pStart = pIn; while (pIn < pInLimit - 1) { BYTE const diff = (*pMatch) ^ *(pIn); if (!diff) { pIn++; pMatch++; continue; } return (U32)(pIn - pStart); } return (U32)(pIn - pStart); } void LDM_outputConfiguration(void) { printf("=====================\n"); printf("Configuration\n"); printf("Window size log: %d\n", LDM_WINDOW_SIZE_LOG); printf("Min match, hash length: %d, %d\n", LDM_MIN_MATCH_LENGTH, LDM_HASH_LENGTH); printf("LDM_MEMORY_USAGE: %d\n", LDM_MEMORY_USAGE); printf("HASH_ONLY_EVERY_LOG: %d\n", HASH_ONLY_EVERY_LOG); printf("LDM_LAG %d\n", LDM_LAG); printf("=====================\n"); } void LDM_readHeader(const void *src, U64 *compressedSize, U64 *decompressedSize) { const BYTE *ip = (const BYTE *)src; *compressedSize = MEM_readLE64(ip); ip += sizeof(U64); *decompressedSize = MEM_readLE64(ip); // ip += sizeof(U64); } void LDM_initializeCCtx(LDM_CCtx *cctx, const void *src, size_t srcSize, void *dst, size_t maxDstSize) { cctx->isize = srcSize; cctx->maxOSize = maxDstSize; cctx->ibase = (const BYTE *)src; cctx->ip = cctx->ibase; cctx->iend = cctx->ibase + srcSize; cctx->ihashLimit = cctx->iend - LDM_HASH_LENGTH; cctx->imatchLimit = cctx->iend - LDM_MIN_MATCH_LENGTH; cctx->obase = (BYTE *)dst; cctx->op = (BYTE *)dst; cctx->anchor = cctx->ibase; memset(&(cctx->stats), 0, sizeof(cctx->stats)); cctx->hashTable = HASH_createTable(LDM_HASHTABLESIZE_U64, cctx->ibase, LDM_MIN_MATCH_LENGTH, LDM_WINDOW_SIZE); cctx->stats.minOffset = UINT_MAX; cctx->stats.windowSizeLog = LDM_WINDOW_SIZE_LOG; cctx->stats.hashTableSizeLog = LDM_MEMORY_USAGE; cctx->lastPosHashed = NULL; cctx->step = 1; // Fixed to be 1 for now. Changing may break things. cctx->nextIp = cctx->ip + cctx->step; cctx->nextPosHashed = 0; cctx->DEBUG_setNextHash = 0; } void LDM_destroyCCtx(LDM_CCtx *cctx) { HASH_destroyTable(cctx->hashTable); } /** * Finds the "best" match. * * Returns 0 if successful and 1 otherwise (i.e. no match can be found * in the remaining input that is long enough). * * matchLength contains the forward length of the match. */ static int LDM_findBestMatch(LDM_CCtx *cctx, const BYTE **match, U32 *matchLength, U32 *backwardMatchLength) { LDM_hashEntry *entry = NULL; cctx->nextIp = cctx->ip + cctx->step; while (entry == NULL) { hash_t h; checksum_t sum; setNextHash(cctx); h = cctx->nextHash; sum = cctx->nextSum; cctx->ip = cctx->nextIp; cctx->nextIp += cctx->step; if (cctx->ip > cctx->imatchLimit) { return 1; } entry = HASH_getBestEntry(cctx->hashTable, h, sum, cctx->ip, cctx->iend, cctx->anchor, matchLength, backwardMatchLength); if (entry != NULL) { *match = entry->offset + cctx->ibase; } putHashOfCurrentPositionFromHash(cctx, h, sum); } setNextHash(cctx); return 0; } void LDM_encodeLiteralLengthAndLiterals( LDM_CCtx *cctx, BYTE *pToken, const U32 literalLength) { /* Encode the literal length. */ if (literalLength >= RUN_MASK) { int len = (int)literalLength - RUN_MASK; *pToken = (RUN_MASK << ML_BITS); for (; len >= 255; len -= 255) { *(cctx->op)++ = 255; } *(cctx->op)++ = (BYTE)len; } else { *pToken = (BYTE)(literalLength << ML_BITS); } /* Encode the literals. */ memcpy(cctx->op, cctx->anchor, literalLength); cctx->op += literalLength; } void LDM_outputBlock(LDM_CCtx *cctx, const U32 literalLength, const U32 offset, const U32 matchLength) { BYTE *pToken = cctx->op++; /* Encode the literal length and literals. */ LDM_encodeLiteralLengthAndLiterals(cctx, pToken, literalLength); /* Encode the offset. */ MEM_write32(cctx->op, offset); cctx->op += LDM_OFFSET_SIZE; /* Encode the match length. */ if (matchLength >= ML_MASK) { unsigned matchLengthRemaining = matchLength; *pToken += ML_MASK; matchLengthRemaining -= ML_MASK; MEM_write32(cctx->op, 0xFFFFFFFF); while (matchLengthRemaining >= 4*0xFF) { cctx->op += 4; MEM_write32(cctx->op, 0xffffffff); matchLengthRemaining -= 4*0xFF; } cctx->op += matchLengthRemaining / 255; *(cctx->op)++ = (BYTE)(matchLengthRemaining % 255); } else { *pToken += (BYTE)(matchLength); } } // TODO: maxDstSize is unused. This function may seg fault when writing // beyond the size of dst, as it does not check maxDstSize. Writing to // a buffer and performing checks is a possible solution. // // This is based upon lz4. size_t LDM_compress(const void *src, size_t srcSize, void *dst, size_t maxDstSize) { LDM_CCtx cctx; const BYTE *match = NULL; U32 forwardMatchLength = 0; U32 backwardsMatchLength = 0; LDM_initializeCCtx(&cctx, src, srcSize, dst, maxDstSize); LDM_outputConfiguration(); /* Hash the first position and put it into the hash table. */ LDM_putHashOfCurrentPosition(&cctx); #if LDM_LAG cctx.lagIp = cctx.ip; cctx.lagHash = cctx.lastHash; cctx.lagSum = cctx.lastSum; #endif /** * Find a match. * If no more matches can be found (i.e. the length of the remaining input * is less than the minimum match length), then stop searching for matches * and encode the final literals. */ while (LDM_findBestMatch(&cctx, &match, &forwardMatchLength, &backwardsMatchLength) == 0) { #ifdef COMPUTE_STATS cctx.stats.numMatches++; #endif cctx.ip -= backwardsMatchLength; match -= backwardsMatchLength; /** * Write current block (literals, literal length, match offset, match * length) and update pointers and hashes. */ { const U32 literalLength = cctx.ip - cctx.anchor; const U32 offset = cctx.ip - match; const U32 matchLength = forwardMatchLength + backwardsMatchLength - LDM_MIN_MATCH_LENGTH; LDM_outputBlock(&cctx, literalLength, offset, matchLength); #ifdef COMPUTE_STATS cctx.stats.totalLiteralLength += literalLength; cctx.stats.totalOffset += offset; cctx.stats.totalMatchLength += matchLength + LDM_MIN_MATCH_LENGTH; cctx.stats.minOffset = offset < cctx.stats.minOffset ? offset : cctx.stats.minOffset; cctx.stats.maxOffset = offset > cctx.stats.maxOffset ? offset : cctx.stats.maxOffset; cctx.stats.offsetHistogram[(U32)intLog2(offset)]++; #endif // Move ip to end of block, inserting hashes at each position. cctx.nextIp = cctx.ip + cctx.step; while (cctx.ip < cctx.anchor + LDM_MIN_MATCH_LENGTH + matchLength + literalLength) { if (cctx.ip > cctx.lastPosHashed) { // TODO: Simplify. LDM_updateLastHashFromNextHash(&cctx); setNextHash(&cctx); } cctx.ip++; cctx.nextIp++; } } // Set start of next block to current input pointer. cctx.anchor = cctx.ip; LDM_updateLastHashFromNextHash(&cctx); } // HASH_outputTableOffsetHistogram(&cctx); /* Encode the last literals (no more matches). */ { const U32 lastRun = cctx.iend - cctx.anchor; BYTE *pToken = cctx.op++; LDM_encodeLiteralLengthAndLiterals(&cctx, pToken, lastRun); } #ifdef COMPUTE_STATS LDM_printCompressStats(&cctx.stats); HASH_outputTableOccupancy(cctx.hashTable); #endif { const size_t ret = cctx.op - cctx.obase; LDM_destroyCCtx(&cctx); return ret; } } struct LDM_DCtx { size_t compressedSize; size_t maxDecompressedSize; const BYTE *ibase; /* Base of input */ const BYTE *ip; /* Current input position */ const BYTE *iend; /* End of source */ const BYTE *obase; /* Base of output */ BYTE *op; /* Current output position */ const BYTE *oend; /* End of output */ }; void LDM_initializeDCtx(LDM_DCtx *dctx, const void *src, size_t compressedSize, void *dst, size_t maxDecompressedSize) { dctx->compressedSize = compressedSize; dctx->maxDecompressedSize = maxDecompressedSize; dctx->ibase = src; dctx->ip = (const BYTE *)src; dctx->iend = dctx->ip + dctx->compressedSize; dctx->op = dst; dctx->oend = dctx->op + dctx->maxDecompressedSize; } size_t LDM_decompress(const void *src, size_t compressedSize, void *dst, size_t maxDecompressedSize) { LDM_DCtx dctx; LDM_initializeDCtx(&dctx, src, compressedSize, dst, maxDecompressedSize); while (dctx.ip < dctx.iend) { BYTE *cpy; const BYTE *match; size_t length, offset; /* Get the literal length. */ const unsigned token = *(dctx.ip)++; if ((length = (token >> ML_BITS)) == RUN_MASK) { unsigned s; do { s = *(dctx.ip)++; length += s; } while (s == 255); } /* Copy the literals. */ cpy = dctx.op + length; memcpy(dctx.op, dctx.ip, length); dctx.ip += length; dctx.op = cpy; //TODO : dynamic offset size offset = MEM_read32(dctx.ip); dctx.ip += LDM_OFFSET_SIZE; match = dctx.op - offset; /* Get the match length. */ length = token & ML_MASK; if (length == ML_MASK) { unsigned s; do { s = *(dctx.ip)++; length += s; } while (s == 255); } length += LDM_MIN_MATCH_LENGTH; /* Copy match. */ cpy = dctx.op + length; // Inefficient for now. while (match < cpy - offset && dctx.op < dctx.oend) { *(dctx.op)++ = *match++; } } return dctx.op - (BYTE *)dst; } // TODO: implement and test hash function void LDM_test(const BYTE *src) { (void)src; }