| /* |
| * Copyright (c) Yann Collet, Facebook, Inc. |
| * All rights reserved. |
| * |
| * This source code is licensed under both the BSD-style license (found in the |
| * LICENSE file in the root directory of this source tree) and the GPLv2 (found |
| * in the COPYING file in the root directory of this source tree). |
| * You may select, at your option, one of the above-listed licenses. |
| */ |
| |
| #ifndef DICTBUILDER_H_001 |
| #define DICTBUILDER_H_001 |
| |
| #if defined (__cplusplus) |
| extern "C" { |
| #endif |
| |
| |
| /*====== Dependencies ======*/ |
| #include <stddef.h> /* size_t */ |
| |
| |
| /* ===== ZDICTLIB_API : control library symbols visibility ===== */ |
| #ifndef ZDICTLIB_VISIBILITY |
| # if defined(__GNUC__) && (__GNUC__ >= 4) |
| # define ZDICTLIB_VISIBILITY __attribute__ ((visibility ("default"))) |
| # else |
| # define ZDICTLIB_VISIBILITY |
| # endif |
| #endif |
| #if defined(ZSTD_DLL_EXPORT) && (ZSTD_DLL_EXPORT==1) |
| # define ZDICTLIB_API __declspec(dllexport) ZDICTLIB_VISIBILITY |
| #elif defined(ZSTD_DLL_IMPORT) && (ZSTD_DLL_IMPORT==1) |
| # define ZDICTLIB_API __declspec(dllimport) ZDICTLIB_VISIBILITY /* It isn't required but allows to generate better code, saving a function pointer load from the IAT and an indirect jump.*/ |
| #else |
| # define ZDICTLIB_API ZDICTLIB_VISIBILITY |
| #endif |
| |
| /******************************************************************************* |
| * Zstd dictionary builder |
| * |
| * FAQ |
| * === |
| * Why should I use a dictionary? |
| * ------------------------------ |
| * |
| * Zstd can use dictionaries to improve compression ratio of small data. |
| * Traditionally small files don't compress well because there is very little |
| * repetion in a single sample, since it is small. But, if you are compressing |
| * many similar files, like a bunch of JSON records that share the same |
| * structure, you can train a dictionary on ahead of time on some samples of |
| * these files. Then, zstd can use the dictionary to find repetitions that are |
| * present across samples. This can vastly improve compression ratio. |
| * |
| * When is a dictionary useful? |
| * ---------------------------- |
| * |
| * Dictionaries are useful when compressing many small files that are similar. |
| * The larger a file is, the less benefit a dictionary will have. Generally, |
| * we don't expect dictionary compression to be effective past 100KB. And the |
| * smaller a file is, the more we would expect the dictionary to help. |
| * |
| * How do I use a dictionary? |
| * -------------------------- |
| * |
| * Simply pass the dictionary to the zstd compressor with |
| * `ZSTD_CCtx_loadDictionary()`. The same dictionary must then be passed to |
| * the decompressor, using `ZSTD_DCtx_loadDictionary()`. There are other |
| * more advanced functions that allow selecting some options, see zstd.h for |
| * complete documentation. |
| * |
| * What is a zstd dictionary? |
| * -------------------------- |
| * |
| * A zstd dictionary has two pieces: Its header, and its content. The header |
| * contains a magic number, the dictionary ID, and entropy tables. These |
| * entropy tables allow zstd to save on header costs in the compressed file, |
| * which really matters for small data. The content is just bytes, which are |
| * repeated content that is common across many samples. |
| * |
| * What is a raw content dictionary? |
| * --------------------------------- |
| * |
| * A raw content dictionary is just bytes. It doesn't have a zstd dictionary |
| * header, a dictionary ID, or entropy tables. Any buffer is a valid raw |
| * content dictionary. |
| * |
| * How do I train a dictionary? |
| * ---------------------------- |
| * |
| * Gather samples from your use case. These samples should be similar to each |
| * other. If you have several use cases, you could try to train one dictionary |
| * per use case. |
| * |
| * Pass those samples to `ZDICT_trainFromBuffer()` and that will train your |
| * dictionary. There are a few advanced versions of this function, but this |
| * is a great starting point. If you want to further tune your dictionary |
| * you could try `ZDICT_optimizeTrainFromBuffer_cover()`. If that is too slow |
| * you can try `ZDICT_optimizeTrainFromBuffer_fastCover()`. |
| * |
| * If the dictionary training function fails, that is likely because you |
| * either passed too few samples, or a dictionary would not be effective |
| * for your data. Look at the messages that the dictionary trainer printed, |
| * if it doesn't say too few samples, then a dictionary would not be effective. |
| * |
| * How large should my dictionary be? |
| * ---------------------------------- |
| * |
| * A reasonable dictionary size, the `dictBufferCapacity`, is about 100KB. |
| * The zstd CLI defaults to a 110KB dictionary. You likely don't need a |
| * dictionary larger than that. But, most use cases can get away with a |
| * smaller dictionary. The advanced dictionary builders can automatically |
| * shrink the dictionary for you, and select a the smallest size that |
| * doesn't hurt compression ratio too much. See the `shrinkDict` parameter. |
| * A smaller dictionary can save memory, and potentially speed up |
| * compression. |
| * |
| * How many samples should I provide to the dictionary builder? |
| * ------------------------------------------------------------ |
| * |
| * We generally recommend passing ~100x the size of the dictionary |
| * in samples. A few thousand should suffice. Having too few samples |
| * can hurt the dictionaries effectiveness. Having more samples will |
| * only improve the dictionaries effectiveness. But having too many |
| * samples can slow down the dictionary builder. |
| * |
| * How do I determine if a dictionary will be effective? |
| * ----------------------------------------------------- |
| * |
| * Simply train a dictionary and try it out. You can use zstd's built in |
| * benchmarking tool to test the dictionary effectiveness. |
| * |
| * # Benchmark levels 1-3 without a dictionary |
| * zstd -b1e3 -r /path/to/my/files |
| * # Benchmark levels 1-3 with a dictioanry |
| * zstd -b1e3 -r /path/to/my/files -D /path/to/my/dictionary |
| * |
| * When should I retrain a dictionary? |
| * ----------------------------------- |
| * |
| * You should retrain a dictionary when its effectiveness drops. Dictionary |
| * effectiveness drops as the data you are compressing changes. Generally, we do |
| * expect dictionaries to "decay" over time, as your data changes, but the rate |
| * at which they decay depends on your use case. Internally, we regularly |
| * retrain dictionaries, and if the new dictionary performs significantly |
| * better than the old dictionary, we will ship the new dictionary. |
| * |
| * I have a raw content dictionary, how do I turn it into a zstd dictionary? |
| * ------------------------------------------------------------------------- |
| * |
| * If you have a raw content dictionary, e.g. by manually constructing it, or |
| * using a third-party dictionary builder, you can turn it into a zstd |
| * dictionary by using `ZDICT_finalizeDictionary()`. You'll also have to |
| * provide some samples of the data. It will add the zstd header to the |
| * raw content, which contains a dictionary ID and entropy tables, which |
| * will improve compression ratio, and allow zstd to write the dictionary ID |
| * into the frame, if you so choose. |
| * |
| * Do I have to use zstd's dictionary builder? |
| * ------------------------------------------- |
| * |
| * No! You can construct dictionary content however you please, it is just |
| * bytes. It will always be valid as a raw content dictionary. If you want |
| * a zstd dictionary, which can improve compression ratio, use |
| * `ZDICT_finalizeDictionary()`. |
| * |
| * What is the attack surface of a zstd dictionary? |
| * ------------------------------------------------ |
| * |
| * Zstd is heavily fuzz tested, including loading fuzzed dictionaries, so |
| * zstd should never crash, or access out-of-bounds memory no matter what |
| * the dictionary is. However, if an attacker can control the dictionary |
| * during decompression, they can cause zstd to generate arbitrary bytes, |
| * just like if they controlled the compressed data. |
| * |
| ******************************************************************************/ |
| |
| |
| /*! ZDICT_trainFromBuffer(): |
| * Train a dictionary from an array of samples. |
| * Redirect towards ZDICT_optimizeTrainFromBuffer_fastCover() single-threaded, with d=8, steps=4, |
| * f=20, and accel=1. |
| * Samples must be stored concatenated in a single flat buffer `samplesBuffer`, |
| * supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order. |
| * The resulting dictionary will be saved into `dictBuffer`. |
| * @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`) |
| * or an error code, which can be tested with ZDICT_isError(). |
| * Note: Dictionary training will fail if there are not enough samples to construct a |
| * dictionary, or if most of the samples are too small (< 8 bytes being the lower limit). |
| * If dictionary training fails, you should use zstd without a dictionary, as the dictionary |
| * would've been ineffective anyways. If you believe your samples would benefit from a dictionary |
| * please open an issue with details, and we can look into it. |
| * Note: ZDICT_trainFromBuffer()'s memory usage is about 6 MB. |
| * Tips: In general, a reasonable dictionary has a size of ~ 100 KB. |
| * It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`. |
| * In general, it's recommended to provide a few thousands samples, though this can vary a lot. |
| * It's recommended that total size of all samples be about ~x100 times the target size of dictionary. |
| */ |
| ZDICTLIB_API size_t ZDICT_trainFromBuffer(void* dictBuffer, size_t dictBufferCapacity, |
| const void* samplesBuffer, |
| const size_t* samplesSizes, unsigned nbSamples); |
| |
| typedef struct { |
| int compressionLevel; /*< optimize for a specific zstd compression level; 0 means default */ |
| unsigned notificationLevel; /*< Write log to stderr; 0 = none (default); 1 = errors; 2 = progression; 3 = details; 4 = debug; */ |
| unsigned dictID; /*< force dictID value; 0 means auto mode (32-bits random value) |
| * NOTE: The zstd format reserves some dictionary IDs for future use. |
| * You may use them in private settings, but be warned that they |
| * may be used by zstd in a public dictionary registry in the future. |
| * These dictionary IDs are: |
| * - low range : <= 32767 |
| * - high range : >= (2^31) |
| */ |
| } ZDICT_params_t; |
| |
| /*! ZDICT_finalizeDictionary(): |
| * Given a custom content as a basis for dictionary, and a set of samples, |
| * finalize dictionary by adding headers and statistics according to the zstd |
| * dictionary format. |
| * |
| * Samples must be stored concatenated in a flat buffer `samplesBuffer`, |
| * supplied with an array of sizes `samplesSizes`, providing the size of each |
| * sample in order. The samples are used to construct the statistics, so they |
| * should be representative of what you will compress with this dictionary. |
| * |
| * The compression level can be set in `parameters`. You should pass the |
| * compression level you expect to use in production. The statistics for each |
| * compression level differ, so tuning the dictionary for the compression level |
| * can help quite a bit. |
| * |
| * You can set an explicit dictionary ID in `parameters`, or allow us to pick |
| * a random dictionary ID for you, but we can't guarantee no collisions. |
| * |
| * The dstDictBuffer and the dictContent may overlap, and the content will be |
| * appended to the end of the header. If the header + the content doesn't fit in |
| * maxDictSize the beginning of the content is truncated to make room, since it |
| * is presumed that the most profitable content is at the end of the dictionary, |
| * since that is the cheapest to reference. |
| * |
| * `dictContentSize` must be >= ZDICT_CONTENTSIZE_MIN bytes. |
| * `maxDictSize` must be >= max(dictContentSize, ZSTD_DICTSIZE_MIN). |
| * |
| * @return: size of dictionary stored into `dstDictBuffer` (<= `maxDictSize`), |
| * or an error code, which can be tested by ZDICT_isError(). |
| * Note: ZDICT_finalizeDictionary() will push notifications into stderr if |
| * instructed to, using notificationLevel>0. |
| * NOTE: This function currently may fail in several edge cases including: |
| * * Not enough samples |
| * * Samples are uncompressible |
| * * Samples are all exactly the same |
| */ |
| ZDICTLIB_API size_t ZDICT_finalizeDictionary(void* dstDictBuffer, size_t maxDictSize, |
| const void* dictContent, size_t dictContentSize, |
| const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples, |
| ZDICT_params_t parameters); |
| |
| |
| /*====== Helper functions ======*/ |
| ZDICTLIB_API unsigned ZDICT_getDictID(const void* dictBuffer, size_t dictSize); /**< extracts dictID; @return zero if error (not a valid dictionary) */ |
| ZDICTLIB_API size_t ZDICT_getDictHeaderSize(const void* dictBuffer, size_t dictSize); /* returns dict header size; returns a ZSTD error code on failure */ |
| ZDICTLIB_API unsigned ZDICT_isError(size_t errorCode); |
| ZDICTLIB_API const char* ZDICT_getErrorName(size_t errorCode); |
| |
| |
| |
| #ifdef ZDICT_STATIC_LINKING_ONLY |
| |
| /* ==================================================================================== |
| * The definitions in this section are considered experimental. |
| * They should never be used with a dynamic library, as they may change in the future. |
| * They are provided for advanced usages. |
| * Use them only in association with static linking. |
| * ==================================================================================== */ |
| |
| #define ZDICT_CONTENTSIZE_MIN 128 |
| #define ZDICT_DICTSIZE_MIN 256 |
| |
| /*! ZDICT_cover_params_t: |
| * k and d are the only required parameters. |
| * For others, value 0 means default. |
| */ |
| typedef struct { |
| unsigned k; /* Segment size : constraint: 0 < k : Reasonable range [16, 2048+] */ |
| unsigned d; /* dmer size : constraint: 0 < d <= k : Reasonable range [6, 16] */ |
| unsigned steps; /* Number of steps : Only used for optimization : 0 means default (40) : Higher means more parameters checked */ |
| unsigned nbThreads; /* Number of threads : constraint: 0 < nbThreads : 1 means single-threaded : Only used for optimization : Ignored if ZSTD_MULTITHREAD is not defined */ |
| double splitPoint; /* Percentage of samples used for training: Only used for optimization : the first nbSamples * splitPoint samples will be used to training, the last nbSamples * (1 - splitPoint) samples will be used for testing, 0 means default (1.0), 1.0 when all samples are used for both training and testing */ |
| unsigned shrinkDict; /* Train dictionaries to shrink in size starting from the minimum size and selects the smallest dictionary that is shrinkDictMaxRegression% worse than the largest dictionary. 0 means no shrinking and 1 means shrinking */ |
| unsigned shrinkDictMaxRegression; /* Sets shrinkDictMaxRegression so that a smaller dictionary can be at worse shrinkDictMaxRegression% worse than the max dict size dictionary. */ |
| ZDICT_params_t zParams; |
| } ZDICT_cover_params_t; |
| |
| typedef struct { |
| unsigned k; /* Segment size : constraint: 0 < k : Reasonable range [16, 2048+] */ |
| unsigned d; /* dmer size : constraint: 0 < d <= k : Reasonable range [6, 16] */ |
| unsigned f; /* log of size of frequency array : constraint: 0 < f <= 31 : 1 means default(20)*/ |
| unsigned steps; /* Number of steps : Only used for optimization : 0 means default (40) : Higher means more parameters checked */ |
| unsigned nbThreads; /* Number of threads : constraint: 0 < nbThreads : 1 means single-threaded : Only used for optimization : Ignored if ZSTD_MULTITHREAD is not defined */ |
| double splitPoint; /* Percentage of samples used for training: Only used for optimization : the first nbSamples * splitPoint samples will be used to training, the last nbSamples * (1 - splitPoint) samples will be used for testing, 0 means default (0.75), 1.0 when all samples are used for both training and testing */ |
| unsigned accel; /* Acceleration level: constraint: 0 < accel <= 10, higher means faster and less accurate, 0 means default(1) */ |
| unsigned shrinkDict; /* Train dictionaries to shrink in size starting from the minimum size and selects the smallest dictionary that is shrinkDictMaxRegression% worse than the largest dictionary. 0 means no shrinking and 1 means shrinking */ |
| unsigned shrinkDictMaxRegression; /* Sets shrinkDictMaxRegression so that a smaller dictionary can be at worse shrinkDictMaxRegression% worse than the max dict size dictionary. */ |
| |
| ZDICT_params_t zParams; |
| } ZDICT_fastCover_params_t; |
| |
| /*! ZDICT_trainFromBuffer_cover(): |
| * Train a dictionary from an array of samples using the COVER algorithm. |
| * Samples must be stored concatenated in a single flat buffer `samplesBuffer`, |
| * supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order. |
| * The resulting dictionary will be saved into `dictBuffer`. |
| * @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`) |
| * or an error code, which can be tested with ZDICT_isError(). |
| * See ZDICT_trainFromBuffer() for details on failure modes. |
| * Note: ZDICT_trainFromBuffer_cover() requires about 9 bytes of memory for each input byte. |
| * Tips: In general, a reasonable dictionary has a size of ~ 100 KB. |
| * It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`. |
| * In general, it's recommended to provide a few thousands samples, though this can vary a lot. |
| * It's recommended that total size of all samples be about ~x100 times the target size of dictionary. |
| */ |
| ZDICTLIB_API size_t ZDICT_trainFromBuffer_cover( |
| void *dictBuffer, size_t dictBufferCapacity, |
| const void *samplesBuffer, const size_t *samplesSizes, unsigned nbSamples, |
| ZDICT_cover_params_t parameters); |
| |
| /*! ZDICT_optimizeTrainFromBuffer_cover(): |
| * The same requirements as above hold for all the parameters except `parameters`. |
| * This function tries many parameter combinations and picks the best parameters. |
| * `*parameters` is filled with the best parameters found, |
| * dictionary constructed with those parameters is stored in `dictBuffer`. |
| * |
| * All of the parameters d, k, steps are optional. |
| * If d is non-zero then we don't check multiple values of d, otherwise we check d = {6, 8}. |
| * if steps is zero it defaults to its default value. |
| * If k is non-zero then we don't check multiple values of k, otherwise we check steps values in [50, 2000]. |
| * |
| * @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`) |
| * or an error code, which can be tested with ZDICT_isError(). |
| * On success `*parameters` contains the parameters selected. |
| * See ZDICT_trainFromBuffer() for details on failure modes. |
| * Note: ZDICT_optimizeTrainFromBuffer_cover() requires about 8 bytes of memory for each input byte and additionally another 5 bytes of memory for each byte of memory for each thread. |
| */ |
| ZDICTLIB_API size_t ZDICT_optimizeTrainFromBuffer_cover( |
| void* dictBuffer, size_t dictBufferCapacity, |
| const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples, |
| ZDICT_cover_params_t* parameters); |
| |
| /*! ZDICT_trainFromBuffer_fastCover(): |
| * Train a dictionary from an array of samples using a modified version of COVER algorithm. |
| * Samples must be stored concatenated in a single flat buffer `samplesBuffer`, |
| * supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order. |
| * d and k are required. |
| * All other parameters are optional, will use default values if not provided |
| * The resulting dictionary will be saved into `dictBuffer`. |
| * @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`) |
| * or an error code, which can be tested with ZDICT_isError(). |
| * See ZDICT_trainFromBuffer() for details on failure modes. |
| * Note: ZDICT_trainFromBuffer_fastCover() requires 6 * 2^f bytes of memory. |
| * Tips: In general, a reasonable dictionary has a size of ~ 100 KB. |
| * It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`. |
| * In general, it's recommended to provide a few thousands samples, though this can vary a lot. |
| * It's recommended that total size of all samples be about ~x100 times the target size of dictionary. |
| */ |
| ZDICTLIB_API size_t ZDICT_trainFromBuffer_fastCover(void *dictBuffer, |
| size_t dictBufferCapacity, const void *samplesBuffer, |
| const size_t *samplesSizes, unsigned nbSamples, |
| ZDICT_fastCover_params_t parameters); |
| |
| /*! ZDICT_optimizeTrainFromBuffer_fastCover(): |
| * The same requirements as above hold for all the parameters except `parameters`. |
| * This function tries many parameter combinations (specifically, k and d combinations) |
| * and picks the best parameters. `*parameters` is filled with the best parameters found, |
| * dictionary constructed with those parameters is stored in `dictBuffer`. |
| * All of the parameters d, k, steps, f, and accel are optional. |
| * If d is non-zero then we don't check multiple values of d, otherwise we check d = {6, 8}. |
| * if steps is zero it defaults to its default value. |
| * If k is non-zero then we don't check multiple values of k, otherwise we check steps values in [50, 2000]. |
| * If f is zero, default value of 20 is used. |
| * If accel is zero, default value of 1 is used. |
| * |
| * @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`) |
| * or an error code, which can be tested with ZDICT_isError(). |
| * On success `*parameters` contains the parameters selected. |
| * See ZDICT_trainFromBuffer() for details on failure modes. |
| * Note: ZDICT_optimizeTrainFromBuffer_fastCover() requires about 6 * 2^f bytes of memory for each thread. |
| */ |
| ZDICTLIB_API size_t ZDICT_optimizeTrainFromBuffer_fastCover(void* dictBuffer, |
| size_t dictBufferCapacity, const void* samplesBuffer, |
| const size_t* samplesSizes, unsigned nbSamples, |
| ZDICT_fastCover_params_t* parameters); |
| |
| typedef struct { |
| unsigned selectivityLevel; /* 0 means default; larger => select more => larger dictionary */ |
| ZDICT_params_t zParams; |
| } ZDICT_legacy_params_t; |
| |
| /*! ZDICT_trainFromBuffer_legacy(): |
| * Train a dictionary from an array of samples. |
| * Samples must be stored concatenated in a single flat buffer `samplesBuffer`, |
| * supplied with an array of sizes `samplesSizes`, providing the size of each sample, in order. |
| * The resulting dictionary will be saved into `dictBuffer`. |
| * `parameters` is optional and can be provided with values set to 0 to mean "default". |
| * @return: size of dictionary stored into `dictBuffer` (<= `dictBufferCapacity`) |
| * or an error code, which can be tested with ZDICT_isError(). |
| * See ZDICT_trainFromBuffer() for details on failure modes. |
| * Tips: In general, a reasonable dictionary has a size of ~ 100 KB. |
| * It's possible to select smaller or larger size, just by specifying `dictBufferCapacity`. |
| * In general, it's recommended to provide a few thousands samples, though this can vary a lot. |
| * It's recommended that total size of all samples be about ~x100 times the target size of dictionary. |
| * Note: ZDICT_trainFromBuffer_legacy() will send notifications into stderr if instructed to, using notificationLevel>0. |
| */ |
| ZDICTLIB_API size_t ZDICT_trainFromBuffer_legacy( |
| void* dictBuffer, size_t dictBufferCapacity, |
| const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples, |
| ZDICT_legacy_params_t parameters); |
| |
| |
| /* Deprecation warnings */ |
| /* It is generally possible to disable deprecation warnings from compiler, |
| for example with -Wno-deprecated-declarations for gcc |
| or _CRT_SECURE_NO_WARNINGS in Visual. |
| Otherwise, it's also possible to manually define ZDICT_DISABLE_DEPRECATE_WARNINGS */ |
| #ifdef ZDICT_DISABLE_DEPRECATE_WARNINGS |
| # define ZDICT_DEPRECATED(message) ZDICTLIB_API /* disable deprecation warnings */ |
| #else |
| # define ZDICT_GCC_VERSION (__GNUC__ * 100 + __GNUC_MINOR__) |
| # if defined (__cplusplus) && (__cplusplus >= 201402) /* C++14 or greater */ |
| # define ZDICT_DEPRECATED(message) [[deprecated(message)]] ZDICTLIB_API |
| # elif defined(__clang__) || (ZDICT_GCC_VERSION >= 405) |
| # define ZDICT_DEPRECATED(message) ZDICTLIB_API __attribute__((deprecated(message))) |
| # elif (ZDICT_GCC_VERSION >= 301) |
| # define ZDICT_DEPRECATED(message) ZDICTLIB_API __attribute__((deprecated)) |
| # elif defined(_MSC_VER) |
| # define ZDICT_DEPRECATED(message) ZDICTLIB_API __declspec(deprecated(message)) |
| # else |
| # pragma message("WARNING: You need to implement ZDICT_DEPRECATED for this compiler") |
| # define ZDICT_DEPRECATED(message) ZDICTLIB_API |
| # endif |
| #endif /* ZDICT_DISABLE_DEPRECATE_WARNINGS */ |
| |
| ZDICT_DEPRECATED("use ZDICT_finalizeDictionary() instead") |
| size_t ZDICT_addEntropyTablesFromBuffer(void* dictBuffer, size_t dictContentSize, size_t dictBufferCapacity, |
| const void* samplesBuffer, const size_t* samplesSizes, unsigned nbSamples); |
| |
| |
| #endif /* ZDICT_STATIC_LINKING_ONLY */ |
| |
| #if defined (__cplusplus) |
| } |
| #endif |
| |
| #endif /* DICTBUILDER_H_001 */ |