/* Cascade the selected strategy down to the next-highest one built into
* this binary. */
-#ifdef ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR
+#ifdef ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR
if (cPar.strategy == ZSTD_btultra2) {
cPar.strategy = ZSTD_btultra;
}
-#endif
-#ifdef ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR
if (cPar.strategy == ZSTD_btultra) {
cPar.strategy = ZSTD_btopt;
}
#endif
#ifdef ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR
NULL,
-#else
- ZSTD_compressBlock_btultra,
-#endif
-#ifdef ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR
NULL
#else
+ ZSTD_compressBlock_btultra,
ZSTD_compressBlock_btultra2
#endif
},
case ZSTD_btultra2:
#if !defined(ZSTD_EXCLUDE_BTLAZY2_BLOCK_COMPRESSOR) \
|| !defined(ZSTD_EXCLUDE_BTOPT_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR)
+ || !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR)
assert(srcSize >= HASH_READ_SIZE);
ZSTD_updateTree(ms, iend-HASH_READ_SIZE, iend);
#endif
#if !defined(ZSTD_EXCLUDE_BTLAZY2_BLOCK_COMPRESSOR) \
|| !defined(ZSTD_EXCLUDE_BTOPT_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR)
+ || !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR)
#define ZSTD_LITFREQ_ADD 2 /* scaling factor for litFreq, so that frequencies adapt faster to new stats */
#define ZSTD_MAX_PRICE (1<<30)
}
#endif
-#if !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR)
+#ifndef ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR
static size_t ZSTD_compressBlock_opt2(
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
const void* src, size_t srcSize, const ZSTD_dictMode_e dictMode)
-#ifndef ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR
+#ifndef ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR
/* ZSTD_initStats_ultra():
* make a first compression pass, just to seed stats with more accurate starting values.
* only works on first block, with no dictionary and no ldm.
DEBUGLOG(5, "ZSTD_compressBlock_btultra (srcSize=%zu)", srcSize);
return ZSTD_compressBlock_opt2(ms, seqStore, rep, src, srcSize, ZSTD_noDict);
}
-#endif
-#ifndef ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR
size_t ZSTD_compressBlock_btultra2(
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
const void* src, size_t srcSize)
#if !defined(ZSTD_EXCLUDE_BTLAZY2_BLOCK_COMPRESSOR) \
|| !defined(ZSTD_EXCLUDE_BTOPT_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR) \
- || !defined(ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR)
+ || !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR)
/* used in ZSTD_loadDictionaryContent() */
void ZSTD_updateTree(ZSTD_matchState_t* ms, const BYTE* ip, const BYTE* iend);
#endif
size_t ZSTD_compressBlock_btultra(
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
void const* src, size_t srcSize);
-#endif
-#ifndef ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR
size_t ZSTD_compressBlock_btultra2(
ZSTD_matchState_t* ms, seqStore_t* seqStore, U32 rep[ZSTD_REP_NUM],
void const* src, size_t srcSize);
&& !defined(ZSTD_EXCLUDE_LAZY2_BLOCK_COMPRESSOR) \
&& !defined(ZSTD_EXCLUDE_BTLAZY2_BLOCK_COMPRESSOR) \
&& !defined(ZSTD_EXCLUDE_BTOPT_BLOCK_COMPRESSOR) \
- && !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR) \
- && !defined(ZSTD_EXCLUDE_BTULTRA2_BLOCK_COMPRESSOR)
+ && !defined(ZSTD_EXCLUDE_BTULTRA_BLOCK_COMPRESSOR)
/* Note : these tests should be replaced by proper regression tests,
* but existing ones do not focus on small data + dictionary + all levels.
*/