rocksdb/util/auto_tune_compressor.cc
Sujit Maharjan 3cc76aae83 Fix nightly build failure because preferred compression type was kNoCompression. (#13739)
Summary:
CostAwareCompressor simply ignores the preferred compression type as compression manager setting takes precedence over the compression type setting. Thus, I am removing the assert statement as it itself is unnecessary for this case.

Pull Request resolved: https://github.com/facebook/rocksdb/pull/13739

Test Plan:
Run nightly build test
```bash
make V=1 J=4 -j4 check
```

Reviewed By: hx235

Differential Revision: D77470932

Pulled By: shubhajeet

fbshipit-source-id: ebb69367d2ffb9bd72432fd04b0cd12ce2d6240a
2025-06-28 09:24:14 -07:00

308 lines
12 KiB
C++

// Copyright (c) Meta Platforms, Inc. and affiliates.
// This source code is licensed under both the GPLv2 (found in the
// COPYING file in the root directory) and Apache 2.0 License
// (found in the LICENSE.Apache file in the root directory).
//
#include "util/auto_tune_compressor.h"
#include "options/options_helper.h"
#include "rocksdb/advanced_compression.h"
#include "test_util/sync_point.h"
#include "util/random.h"
#include "util/stop_watch.h"
namespace ROCKSDB_NAMESPACE {
const std::vector<std::vector<int>> CostAwareCompressor::kCompressionLevels{
{0}, // KSnappyCompression
{}, // kZlibCompression
{}, // kBZip2Compression
{1, 4, 9}, // kLZ4Compression
{1, 4, 9}, // klZ4HCCompression
{}, // kXpressCompression
{1, 15, 22} // kZSTD
};
int CompressionRejectionProbabilityPredictor::Predict() const {
return pred_rejection_prob_percentage_;
}
size_t CompressionRejectionProbabilityPredictor::attempted_compression_count()
const {
return rejected_count_ + compressed_count_;
}
bool CompressionRejectionProbabilityPredictor::Record(
Slice uncompressed_block_data, std::string* compressed_output,
const CompressionOptions& opts) {
if (compressed_output->size() >
(static_cast<uint64_t>(opts.max_compressed_bytes_per_kb) *
uncompressed_block_data.size()) >>
10) {
rejected_count_++;
} else {
compressed_count_++;
}
auto attempted = attempted_compression_count();
if (attempted >= window_size_) {
pred_rejection_prob_percentage_ =
static_cast<int>(rejected_count_ * 100 / attempted);
compressed_count_ = 0;
rejected_count_ = 0;
assert(attempted_compression_count() == 0);
}
return true;
}
AutoSkipCompressorWrapper::AutoSkipCompressorWrapper(
std::unique_ptr<Compressor> compressor, const CompressionOptions& opts)
: CompressorWrapper::CompressorWrapper(std::move(compressor)),
opts_(opts) {}
const char* AutoSkipCompressorWrapper::Name() const {
return "AutoSkipCompressorWrapper";
}
Status AutoSkipCompressorWrapper::CompressBlock(
Slice uncompressed_data, std::string* compressed_output,
CompressionType* out_compression_type, ManagedWorkingArea* wa) {
// Check if the managed working area is provided or owned by this object.
// If not, bypass auto-skip logic since the working area lacks a predictor to
// record or make necessary decisions to compress or bypass compression of the
// block
if (wa == nullptr || wa->owner() != this) {
return wrapped_->CompressBlock(uncompressed_data, compressed_output,
out_compression_type, wa);
}
bool exploration =
Random::GetTLSInstance()->PercentTrue(kExplorationPercentage);
TEST_SYNC_POINT_CALLBACK(
"AutoSkipCompressorWrapper::CompressBlock::exploitOrExplore",
&exploration);
auto autoskip_wa = static_cast<AutoSkipWorkingArea*>(wa->get());
if (exploration) {
return CompressBlockAndRecord(uncompressed_data, compressed_output,
out_compression_type, autoskip_wa);
} else {
auto predictor_ptr = autoskip_wa->predictor;
auto prediction = predictor_ptr->Predict();
if (prediction <= kProbabilityCutOff) {
// decide to compress
return CompressBlockAndRecord(uncompressed_data, compressed_output,
out_compression_type, autoskip_wa);
} else {
// decide to bypass compression
*out_compression_type = kNoCompression;
return Status::OK();
}
}
return Status::OK();
}
Compressor::ManagedWorkingArea AutoSkipCompressorWrapper::ObtainWorkingArea() {
auto wrap_wa = wrapped_->ObtainWorkingArea();
return ManagedWorkingArea(new AutoSkipWorkingArea(std::move(wrap_wa)), this);
}
void AutoSkipCompressorWrapper::ReleaseWorkingArea(WorkingArea* wa) {
delete static_cast<AutoSkipWorkingArea*>(wa);
}
Status AutoSkipCompressorWrapper::CompressBlockAndRecord(
Slice uncompressed_data, std::string* compressed_output,
CompressionType* out_compression_type, AutoSkipWorkingArea* wa) {
Status status = wrapped_->CompressBlock(uncompressed_data, compressed_output,
out_compression_type, &(wa->wrapped));
// determine if it was rejected or compressed
auto predictor_ptr = wa->predictor;
predictor_ptr->Record(uncompressed_data, compressed_output, opts_);
return status;
}
const char* AutoSkipCompressorManager::Name() const {
// should have returned "AutoSkipCompressorManager" but we currently have an
// error so for now returning name of the wrapped container
return wrapped_->Name();
}
std::unique_ptr<Compressor> AutoSkipCompressorManager::GetCompressorForSST(
const FilterBuildingContext& context, const CompressionOptions& opts,
CompressionType preferred) {
assert(GetSupportedCompressions().size() > 1);
assert(preferred != kNoCompression);
return std::make_unique<AutoSkipCompressorWrapper>(
wrapped_->GetCompressorForSST(context, opts, preferred), opts);
}
CostAwareCompressor::CostAwareCompressor(const CompressionOptions& opts)
: opts_(opts) {
// Creates compressor supporting all the compression types and levels as per
// the compression levels set in vector CompressionLevels
auto builtInManager = GetBuiltinV2CompressionManager();
const auto& compressions = GetSupportedCompressions();
for (size_t i = 0; i < kCompressionLevels.size(); i++) {
CompressionType type = static_cast<CompressionType>(i + 1);
if (type == kNoCompression) {
continue;
}
if (kCompressionLevels[type - 1].size() == 0) {
allcompressors_.emplace_back();
continue;
} else {
// if the compression type is not supported, then skip and remove
// compression levels from the supported compression level list
if (std::find(compressions.begin(), compressions.end(), type) ==
compressions.end()) {
allcompressors_.emplace_back();
continue;
}
std::vector<std::unique_ptr<Compressor>> compressors_diff_levels;
for (size_t j = 0; j < kCompressionLevels[type - 1].size(); j++) {
auto level = kCompressionLevels[type - 1][j];
CompressionOptions new_opts = opts;
new_opts.level = level;
compressors_diff_levels.push_back(
builtInManager->GetCompressor(new_opts, type));
allcompressors_index_.emplace_back(i, j);
}
allcompressors_.push_back(std::move(compressors_diff_levels));
}
}
}
const char* CostAwareCompressor::Name() const { return "CostAwareCompressor"; }
size_t CostAwareCompressor::GetMaxSampleSizeIfWantDict(
CacheEntryRole block_type) const {
auto idx = allcompressors_index_.back();
return allcompressors_[idx.first][idx.second]->GetMaxSampleSizeIfWantDict(
block_type);
}
Slice CostAwareCompressor::GetSerializedDict() const {
auto idx = allcompressors_index_.back();
return allcompressors_[idx.first][idx.second]->GetSerializedDict();
}
CompressionType CostAwareCompressor::GetPreferredCompressionType() const {
return kZSTD;
}
std::unique_ptr<Compressor> CostAwareCompressor::MaybeCloneSpecialized(
CacheEntryRole block_type, DictSampleArgs&& dict_samples) {
// TODO: full dictionary compression support. Currently this just falls
// back on a non-multi compressor when asked to use a dictionary.
auto idx = allcompressors_index_.back();
return allcompressors_[idx.first][idx.second]->MaybeCloneSpecialized(
block_type, std::move(dict_samples));
}
Status CostAwareCompressor::CompressBlock(Slice uncompressed_data,
std::string* compressed_output,
CompressionType* out_compression_type,
ManagedWorkingArea* wa) {
// Check if the managed working area is provided or owned by this object.
// If not, bypass compressor logic since the working area lacks a predictor
if (allcompressors_.size() == 0) {
return Status::NotSupported("No compression type supported");
}
if (wa == nullptr || wa->owner() != this) {
// highest compression level of Zstd
size_t choosen_compression_type = 6;
size_t compression_level_ptr = 2;
return allcompressors_[choosen_compression_type][compression_level_ptr]
->CompressBlock(uncompressed_data, compressed_output,
out_compression_type, wa);
}
auto local_wa = static_cast<CostAwareWorkingArea*>(wa->get());
std::pair<size_t, size_t> choosen_index(6, 2);
size_t choosen_compression_type = choosen_index.first;
size_t compresion_level_ptr = choosen_index.second;
return CompressBlockAndRecord(choosen_compression_type, compresion_level_ptr,
uncompressed_data, compressed_output,
out_compression_type, local_wa);
}
Compressor::ManagedWorkingArea CostAwareCompressor::ObtainWorkingArea() {
auto wrap_wa = allcompressors_.back().back()->ObtainWorkingArea();
auto wa = new CostAwareWorkingArea(std::move(wrap_wa));
// Create cost predictors for each compression type and level
wa->cost_predictors_.reserve(allcompressors_.size());
for (size_t i = 0; i < allcompressors_.size(); i++) {
CompressionType type = static_cast<CompressionType>(i + 1);
if (allcompressors_[type - 1].size() == 0) {
wa->cost_predictors_.emplace_back();
continue;
} else {
std::vector<IOCPUCostPredictor*> predictors_diff_levels;
predictors_diff_levels.reserve(kCompressionLevels[type - 1].size());
for (size_t j = 0; j < kCompressionLevels[type - 1].size(); j++) {
predictors_diff_levels.emplace_back(new IOCPUCostPredictor(10));
}
wa->cost_predictors_.emplace_back(std::move(predictors_diff_levels));
}
}
return ManagedWorkingArea(wa, this);
}
void CostAwareCompressor::ReleaseWorkingArea(WorkingArea* wa) {
// remove all created cost predictors
for (auto& prdictors_diff_levels :
static_cast<CostAwareWorkingArea*>(wa)->cost_predictors_) {
for (auto& predictor : prdictors_diff_levels) {
delete predictor;
}
}
delete static_cast<CostAwareWorkingArea*>(wa);
}
Status CostAwareCompressor::CompressBlockAndRecord(
size_t choosen_compression_type, size_t compression_level_ptr,
Slice uncompressed_data, std::string* compressed_output,
CompressionType* out_compression_type, CostAwareWorkingArea* wa) {
assert(choosen_compression_type < allcompressors_.size());
assert(compression_level_ptr <
allcompressors_[choosen_compression_type].size());
assert(choosen_compression_type < wa->cost_predictors_.size());
assert(compression_level_ptr <
wa->cost_predictors_[choosen_compression_type].size());
StopWatchNano<> timer(Env::Default()->GetSystemClock().get(), true);
Status status =
allcompressors_[choosen_compression_type][compression_level_ptr]
->CompressBlock(uncompressed_data, compressed_output,
out_compression_type, &(wa->wrapped_));
std::pair<size_t, size_t> measured_data(timer.ElapsedMicros(),
compressed_output->size());
auto predictor =
wa->cost_predictors_[choosen_compression_type][compression_level_ptr];
auto output_length = measured_data.second;
auto cpu_time = measured_data.first;
predictor->CPUPredictor.Record(cpu_time);
predictor->IOPredictor.Record(output_length);
TEST_SYNC_POINT_CALLBACK(
"CostAwareCompressor::CompressBlockAndRecord::GetPredictor",
wa->cost_predictors_[choosen_compression_type][compression_level_ptr]);
return status;
}
std::shared_ptr<CompressionManagerWrapper> CreateAutoSkipCompressionManager(
std::shared_ptr<CompressionManager> wrapped) {
return std::make_shared<AutoSkipCompressorManager>(
wrapped == nullptr ? GetBuiltinV2CompressionManager() : wrapped);
}
const char* CostAwareCompressorManager::Name() const {
// should have returned "CostAwareCompressorManager" but we currently have an
// error so for now returning name of the wrapped container
return wrapped_->Name();
}
std::unique_ptr<Compressor> CostAwareCompressorManager::GetCompressorForSST(
const FilterBuildingContext& context, const CompressionOptions& opts,
CompressionType preferred) {
assert(GetSupportedCompressions().size() > 1);
(void)context;
(void)preferred;
return std::make_unique<CostAwareCompressor>(opts);
}
std::shared_ptr<CompressionManagerWrapper> CreateCostAwareCompressionManager(
std::shared_ptr<CompressionManager> wrapped) {
return std::make_shared<CostAwareCompressorManager>(
wrapped == nullptr ? GetBuiltinV2CompressionManager() : wrapped);
}
} // namespace ROCKSDB_NAMESPACE