/dragonfly/contrib/zstd/lib/compress/ |
H A D | zstd_compress_literals.c | 19 U32 const flSize = 1 + (srcSize>31) + (srcSize>4095); in ZSTD_noCompressLiterals() 26 ostart[0] = (BYTE)((U32)set_basic + (srcSize<<3)); in ZSTD_noCompressLiterals() 38 ZSTD_memcpy(ostart + flSize, src, srcSize); in ZSTD_noCompressLiterals() 39 DEBUGLOG(5, "Raw literals: %u -> %u", (U32)srcSize, (U32)(srcSize + flSize)); in ZSTD_noCompressLiterals() 40 return srcSize + flSize; in ZSTD_noCompressLiterals() 46 U32 const flSize = 1 + (srcSize>31) + (srcSize>4095); in ZSTD_compressRleLiteralsBlock() 74 const void* src, size_t srcSize, in ZSTD_compressLiterals() argument 79 size_t const lhSize = 3 + (srcSize >= 1 KB) + (srcSize >= 16 KB); in ZSTD_compressLiterals() 81 U32 singleStream = srcSize < 256; in ZSTD_compressLiterals() 86 disableLiteralCompression, (U32)srcSize); in ZSTD_compressLiterals() [all …]
|
H A D | zstd_lazy.h | 36 void const* src, size_t srcSize); 39 void const* src, size_t srcSize); 42 void const* src, size_t srcSize); 45 void const* src, size_t srcSize); 49 void const* src, size_t srcSize); 52 void const* src, size_t srcSize); 55 void const* src, size_t srcSize); 58 void const* src, size_t srcSize); 62 void const* src, size_t srcSize); 65 void const* src, size_t srcSize); [all …]
|
H A D | huf_compress.c | 577 n = srcSize & ~3; /* join to mod 4 */ in HUF_compress1X_usingCTable_internal_body() 578 switch (srcSize & 3) in HUF_compress1X_usingCTable_internal_body() 662 const BYTE* const iend = ip + srcSize; in HUF_compress4X_usingCTable_internal() 717 const void* src, size_t srcSize, in HUF_compressCTable_internal() argument 743 const void* src, size_t srcSize, in HUF_compress_internal() argument 760 if (!srcSize) return 0; /* Uncompressed */ in HUF_compress_internal() 819 if (hSize + 12ul >= srcSize) { return 0; } in HUF_compress_internal() 826 src, srcSize, in HUF_compress_internal() 832 const void* src, size_t srcSize, in HUF_compress1X_wksp() argument 895 const void* src, size_t srcSize, in HUF_compress1X() argument [all …]
|
H A D | fse_compress.c | 322 U32 minBitsSrc = BIT_highbit32((U32)(srcSize)) + 1; in FSE_minTableLog() 333 U32 minBits = FSE_minTableLog(srcSize, maxSymbolValue); in FSE_optimalTableLog_internal() 558 const void* src, size_t srcSize, in FSE_compress_usingCTable_generic() argument 562 const BYTE* const iend = istart + srcSize; in FSE_compress_usingCTable_generic() 569 if (srcSize <= 2) return 0; in FSE_compress_usingCTable_generic() 575 if (srcSize & 1) { in FSE_compress_usingCTable_generic() 586 srcSize -= 2; in FSE_compress_usingCTable_generic() 617 const void* src, size_t srcSize, in FSE_compress_usingCTable() argument 651 if (srcSize <= 1) return 0; /* Not compressible */ in FSE_compress_wksp() 663 …ECK_F( FSE_normalizeCount(norm, tableLog, count, srcSize, maxSymbolValue, /* useLowProbCount */ sr… in FSE_compress_wksp() [all …]
|
H A D | zstd_opt.h | 25 void const* src, size_t srcSize); 28 void const* src, size_t srcSize); 31 void const* src, size_t srcSize); 36 void const* src, size_t srcSize); 39 void const* src, size_t srcSize); 43 void const* src, size_t srcSize); 46 void const* src, size_t srcSize);
|
H A D | zstd_fast.c | 49 void const* src, size_t srcSize, in ZSTD_compressBlock_fast_generic() argument 63 const U32 endIndex = (U32)((size_t)(istart - base) + srcSize); in ZSTD_compressBlock_fast_generic() 66 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_fast_generic() 188 void const* src, size_t srcSize) in ZSTD_compressBlock_fast() argument 209 void const* src, size_t srcSize, U32 const mls) in ZSTD_compressBlock_fast_dictMatchState_generic() argument 222 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_fast_dictMatchState_generic() 241 const U32 endIndex = (U32)((size_t)(ip - base) + srcSize); in ZSTD_compressBlock_fast_dictMatchState_generic() 356 void const* src, size_t srcSize) in ZSTD_compressBlock_fast_dictMatchState() argument 377 void const* src, size_t srcSize, U32 const mls) in ZSTD_compressBlock_fast_extDict_generic() argument 397 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_fast_extDict_generic() [all …]
|
H A D | hist.h | 31 const void* src, size_t srcSize); 47 const void* src, size_t srcSize, 55 const void* src, size_t srcSize); 63 const void* src, size_t srcSize, 75 const void* src, size_t srcSize);
|
H A D | zstd_ldm.c | 292 BYTE const* const iend = istart + srcSize; in ZSTD_ldm_generateSequences_internal() 450 BYTE const* const iend = istart + srcSize; in ZSTD_ldm_generateSequences() 452 size_t const nbChunks = (srcSize / kMaxChunkSize) + ((srcSize % kMaxChunkSize) != 0); in ZSTD_ldm_generateSequences() 524 if (srcSize <= seq->litLength) { in ZSTD_ldm_skipSequences() 526 seq->litLength -= (U32)srcSize; in ZSTD_ldm_skipSequences() 529 srcSize -= seq->litLength; in ZSTD_ldm_skipSequences() 531 if (srcSize < seq->matchLength) { in ZSTD_ldm_skipSequences() 533 seq->matchLength -= (U32)srcSize; in ZSTD_ldm_skipSequences() 543 srcSize -= seq->matchLength; in ZSTD_ldm_skipSequences() 599 void const* src, size_t srcSize) in ZSTD_ldm_blockCompress() argument [all …]
|
H A D | zstd_double_fast.c | 53 void const* src, size_t srcSize, in ZSTD_compressBlock_doubleFast_generic() argument 65 const U32 endIndex = (U32)((size_t)(istart - base) + srcSize); in ZSTD_compressBlock_doubleFast_generic() 69 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_doubleFast_generic() 321 void const* src, size_t srcSize) in ZSTD_compressBlock_doubleFast() argument 341 void const* src, size_t srcSize) in ZSTD_compressBlock_doubleFast_dictMatchState() argument 361 void const* src, size_t srcSize, in ZSTD_compressBlock_doubleFast_extDict_generic() argument 372 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_doubleFast_extDict_generic() 375 const U32 endIndex = (U32)((size_t)(istart - base) + srcSize); in ZSTD_compressBlock_doubleFast_extDict_generic() 386 DEBUGLOG(5, "ZSTD_compressBlock_doubleFast_extDict_generic (srcSize=%zu)", srcSize); in ZSTD_compressBlock_doubleFast_extDict_generic() 506 void const* src, size_t srcSize) in ZSTD_compressBlock_doubleFast_extDict() argument [all …]
|
H A D | zstd_compress.c | 56 size_t ZSTD_compressBound(size_t srcSize) { in ZSTD_compressBound() argument 57 return ZSTD_COMPRESSBOUND(srcSize); in ZSTD_compressBound() 1192 srcSize = minSrcSize; in ZSTD_adjustCParams_internal() 1208 if ( (srcSize < maxWindowResize) in ZSTD_adjustCParams_internal() 1235 if (srcSize == 0) srcSize = ZSTD_CONTENTSIZE_UNKNOWN; in ZSTD_adjustCParams() 2363 { size_t const maxCSize = srcSize - ZSTD_minGain(srcSize, cctxParams->cParams.strategy); in ZSTD_entropyCompressSequences() 2695 srcSize, in ZSTD_compressBlock_internal() 2712 ZSTD_isRLE(ip, srcSize)) in ZSTD_compressBlock_internal() 2770 … size_t const maxCSize = srcSize - ZSTD_minGain(srcSize, zc->appliedParams.cParams.strategy); in ZSTD_compressBlock_targetCBlockSize_body() 2844 size_t remaining = srcSize; in ZSTD_compress_frameChunk() [all …]
|
H A D | zstd_lazy.c | 1125 void const* src, size_t srcSize) in ZSTD_compressBlock_btlazy2() argument 1132 void const* src, size_t srcSize) in ZSTD_compressBlock_lazy2() argument 1139 void const* src, size_t srcSize) in ZSTD_compressBlock_lazy() argument 1146 void const* src, size_t srcSize) in ZSTD_compressBlock_greedy() argument 1153 void const* src, size_t srcSize) in ZSTD_compressBlock_btlazy2_dictMatchState() argument 1160 void const* src, size_t srcSize) in ZSTD_compressBlock_lazy2_dictMatchState() argument 1167 void const* src, size_t srcSize) in ZSTD_compressBlock_lazy_dictMatchState() argument 1174 void const* src, size_t srcSize) in ZSTD_compressBlock_greedy_dictMatchState() argument 1182 void const* src, size_t srcSize) in ZSTD_compressBlock_lazy2_dedicatedDictSearch() argument 1189 void const* src, size_t srcSize) in ZSTD_compressBlock_lazy_dedicatedDictSearch() argument [all …]
|
H A D | zstd_opt.c | 1216 const void* src, size_t srcSize) in ZSTD_compressBlock_btopt() argument 1270 ms->window.base -= srcSize; in ZSTD_initStats_ultra() 1271 ms->window.dictLimit += (U32)srcSize; in ZSTD_initStats_ultra() 1281 const void* src, size_t srcSize) in ZSTD_compressBlock_btultra() argument 1289 const void* src, size_t srcSize) in ZSTD_compressBlock_btultra2() argument 1302 assert(srcSize <= ZSTD_BLOCKSIZE_MAX); in ZSTD_compressBlock_btultra2() 1307 && (srcSize > ZSTD_PREDEF_THRESHOLD) in ZSTD_compressBlock_btultra2() 1317 const void* src, size_t srcSize) in ZSTD_compressBlock_btopt_dictMatchState() argument 1324 const void* src, size_t srcSize) in ZSTD_compressBlock_btultra_dictMatchState() argument 1331 const void* src, size_t srcSize) in ZSTD_compressBlock_btopt_extDict() argument [all …]
|
H A D | zstd_compress_literals.h | 17 size_t ZSTD_noCompressLiterals (void* dst, size_t dstCapacity, const void* src, size_t srcSize); 19 …e_t ZSTD_compressRleLiteralsBlock (void* dst, size_t dstCapacity, const void* src, size_t srcSize); 25 const void* src, size_t srcSize,
|
H A D | zstd_fast.h | 25 void const* src, size_t srcSize); 28 void const* src, size_t srcSize); 31 void const* src, size_t srcSize);
|
H A D | zstd_double_fast.h | 25 void const* src, size_t srcSize); 28 void const* src, size_t srcSize); 31 void const* src, size_t srcSize);
|
H A D | zstd_ldm.h | 47 ldmParams_t const* params, void const* src, size_t srcSize); 69 void const* src, size_t srcSize); 78 void ZSTD_ldm_skipSequences(rawSeqStore_t* rawSeqStore, size_t srcSize,
|
H A D | hist.c | 30 const void* src, size_t srcSize) in HIST_count_simple() argument 33 const BYTE* const end = ip + srcSize; in HIST_count_simple() 38 if (srcSize==0) { *maxSymbolValuePtr = 0; return 0; } in HIST_count_simple() 176 const void* src, size_t srcSize) in HIST_count() argument 179 … return HIST_count_wksp(count, maxSymbolValuePtr, src, srcSize, tmpCounters, sizeof(tmpCounters)); in HIST_count()
|
H A D | zstd_compress_internal.h | 357 void const* src, size_t srcSize); 435 RETURN_ERROR_IF(srcSize + ZSTD_blockHeaderSize > dstCapacity, in ZSTD_noCompressBlock() 438 ZSTD_memcpy((BYTE*)dst + ZSTD_blockHeaderSize, src, srcSize); in ZSTD_noCompressBlock() 439 return ZSTD_blockHeaderSize + srcSize; in ZSTD_noCompressBlock() 457 MEM_STATIC size_t ZSTD_minGain(size_t srcSize, ZSTD_strategy strat) in ZSTD_minGain() argument 462 return (srcSize >> minlog) + 2; in ZSTD_minGain() 1010 void const* src, size_t srcSize) in ZSTD_window_update() argument 1015 if (srcSize == 0) in ZSTD_window_update() 1033 window->nextSrc = ip + srcSize; in ZSTD_window_update() 1035 if ( (ip+srcSize > window->dictBase + window->lowLimit) in ZSTD_window_update() [all …]
|
/dragonfly/contrib/zstd/lib/decompress/ |
H A D | zstd_decompress.c | 348 if (ZSTD_isLegacy(src, srcSize)) { in ZSTD_getFrameContentSize() 400 srcSize -= skippableSize; in ZSTD_findDecompressedSize() 417 srcSize -= frameSrcSize; in ZSTD_findDecompressedSize() 476 if (ZSTD_isLegacy(src, srcSize)) in ZSTD_findFrameSizeInfo() 557 while (srcSize > 0) { in ZSTD_decompressBound() 565 srcSize -= compressedSize; in ZSTD_decompressBound() 593 if (srcSize == 0) return 0; in ZSTD_copyRawBlock() 596 ZSTD_memcpy(dst, src, srcSize); in ZSTD_copyRawBlock() 597 return srcSize; in ZSTD_copyRawBlock() 740 srcSize -= frameSize; in ZSTD_decompressMultiFrame() [all …]
|
H A D | zstd_decompress_block.c | 56 size_t ZSTD_getcBlockSize(const void* src, size_t srcSize, in ZSTD_getcBlockSize() argument 59 RETURN_ERROR_IF(srcSize < ZSTD_blockHeaderSize, srcSize_wrong, ""); in ZSTD_getcBlockSize() 75 const void* src, size_t srcSize); 531 const void* src, size_t srcSize, in ZSTD_buildSeqTable() argument 540 RETURN_ERROR_IF(!srcSize, srcSize_wrong, ""); in ZSTD_buildSeqTable() 578 const void* src, size_t srcSize) in ZSTD_decodeSeqHeaders() argument 581 const BYTE* const iend = istart + srcSize; in ZSTD_decodeSeqHeaders() 593 RETURN_ERROR_IF(srcSize != 1, srcSize_wrong, ""); in ZSTD_decodeSeqHeaders() 1470 srcSize -= litCSize; in ZSTD_decompressBlock_internal() 1487 srcSize -= seqHSize; in ZSTD_decompressBlock_internal() [all …]
|
/dragonfly/contrib/zstd/lib/ |
H A D | zstd.h | 122 const void* src, size_t srcSize, 179 #define ZSTD_COMPRESSBOUND(srcSize) ((srcSize) + ((srcSize)>>8) + (((srcSize) < (128<<10)) ? (((1… argument 214 const void* src, size_t srcSize, 234 const void* src, size_t srcSize); 519 const void* src, size_t srcSize); 818 const void* src, size_t srcSize, 830 const void* src, size_t srcSize, 865 const void* src, size_t srcSize, 1372 const void* src, size_t srcSize); 1578 const void* src, size_t srcSize, [all …]
|
/dragonfly/contrib/zstd/lib/common/ |
H A D | huf.h | 55 const void* src, size_t srcSize); 87 const void* src, size_t srcSize, 96 const void* src, size_t srcSize, 192 unsigned HUF_optimalTableLog(unsigned maxTableLog, size_t srcSize, unsigned maxSymbolValue); 210 const void* src, size_t srcSize, 232 const void* src, size_t srcSize); 243 const void* src, size_t srcSize, 285 size_t HUF_readDTableX1 (HUF_DTable* DTable, const void* src, size_t srcSize); 286 size_t HUF_readDTableX1_wksp (HUF_DTable* DTable, const void* src, size_t srcSize, void* workSpace,… 289 size_t HUF_readDTableX2 (HUF_DTable* DTable, const void* src, size_t srcSize); [all …]
|
H A D | entropy_common.c | 258 const void* src, size_t srcSize) in HUF_readStats() argument 267 const void* src, size_t srcSize, in HUF_readStats_body() argument 276 if (!srcSize) return ERROR(srcSize_wrong); in HUF_readStats_body() 283 if (iSize+1 > srcSize) return ERROR(srcSize_wrong); in HUF_readStats_body() 292 if (iSize+1 > srcSize) return ERROR(srcSize_wrong); in HUF_readStats_body() 333 const void* src, size_t srcSize, in HUF_readStats_body_default() argument 336 …tats_body(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, workSpace, wkspS… in HUF_readStats_body_default() 342 const void* src, size_t srcSize, in HUF_readStats_body_bmi2() argument 345 …tats_body(huffWeight, hwSize, rankStats, nbSymbolsPtr, tableLogPtr, src, srcSize, workSpace, wkspS… in HUF_readStats_body_bmi2() 351 const void* src, size_t srcSize, in HUF_readStats_wksp() argument [all …]
|
H A D | bitstream.h | 106 MEM_STATIC size_t BIT_initDStream(BIT_DStream_t* bitD, const void* srcBuffer, size_t srcSize); 277 MEM_STATIC size_t BIT_initDStream(BIT_DStream_t* bitD, const void* srcBuffer, size_t srcSize) in BIT_initDStream() argument 279 if (srcSize < 1) { ZSTD_memset(bitD, 0, sizeof(*bitD)); return ERROR(srcSize_wrong); } in BIT_initDStream() 284 if (srcSize >= sizeof(bitD->bitContainer)) { /* normal case */ in BIT_initDStream() 285 bitD->ptr = (const char*)srcBuffer + srcSize - sizeof(bitD->bitContainer); in BIT_initDStream() 287 { BYTE const lastByte = ((const BYTE*)srcBuffer)[srcSize-1]; in BIT_initDStream() 293 switch(srcSize) in BIT_initDStream() 315 { BYTE const lastByte = ((const BYTE*)srcBuffer)[srcSize-1]; in BIT_initDStream() 319 bitD->bitsConsumed += (U32)(sizeof(bitD->bitContainer) - srcSize)*8; in BIT_initDStream() 322 return srcSize; in BIT_initDStream()
|
/dragonfly/contrib/zstd/programs/ |
H A D | benchzstd.c | 151 size_t srcSize; member 231 const void* srcBuffer, size_t srcSize, in local_defaultCompress() argument 249 in.src = srcBuffer; in.size = srcSize; in.pos = 0; in local_defaultDecompress() 319 size_t const loadedCompressedSize = srcSize; in BMK_benchMemAdvancedNoAlloc() 350 cSize = srcSize; in BMK_benchMemAdvancedNoAlloc() 351 srcSize = decodedSize; in BMK_benchMemAdvancedNoAlloc() 352 ratio = (double)srcSize / (double)cSize; in BMK_benchMemAdvancedNoAlloc() 451 ratio = (double)srcSize / cSize; in BMK_benchMemAdvancedNoAlloc() 502 for (u=0; u<srcSize; u++) { in BMK_benchMemAdvancedNoAlloc() 595 void* resultBuffer = srcSize ? malloc(srcSize) : NULL; in BMK_benchMemAdvanced() [all …]
|