Summary: Add file size validation in ReadFooterFromFile function. Deprecate skip_checking_sst_file_sizes_on_db_open option. This change is used to address this issue https://github.com/facebook/rocksdb/issues/13619 It supports file size validation in ReadFooterFromFile. In favor of this change, CheckConsistency function and skip_checking_sst_file_sizes_on_db_open flag are deprecated. The CheckConsistency function checks each file size matches what was recorded in manifest during DB open. Meantime, ReadFooterFromFile was called for each file in LoadTables function. Since ReadFooterFromFile always validates file size, the CheckConsistency is redundant. In addtion, CheckConsistency is executed in a single thread. This could slow down DB open when a network file system is used. Therefore, the flag skip_checking_sst_file_sizes_on_db_open was added to skip this check. After this change, ReadFooterFromFile was executed in parallel through multiple threads. Therefore, the concern of DB open slowness is eliminated, and the flag could be deprecated. When paranoid check flag is set to true, corrupted file will fail to open the DB. When paranoid check flag is set to false, DB will still be able to open, the healthy ones can be accessed, while the corrupted ones not. There is 2 slight concerns of this change. *If max_open_files is set with smaller value, engine will not open all the files during DB open. This means if there is a corruption on file size, it will not be detected during DB open, but rather at a later time. Since the default is -1, which means open all the files, and it is rarely overridden and a lot of new features rely on it to be -1, the risk is very low. *If FIFO compaction is used, engine could fail to open DB unnecessarily on the corrupted files that would never be used again. However, this is a very rare case as well. The error could still be ignored by setting paranoid_checks operationally. The risk is very low. To remain backward compatibility. The public facing flag was kept and marked as no-op internally. Another change is required to fully remove the flag. Pull Request resolved: https://github.com/facebook/rocksdb/pull/13676 Test Plan: make check A new unit test was added to validate file size check API works as expected. Reviewed By: pdillinger Differential Revision: D76168033 Pulled By: xingbowang fbshipit-source-id: 8ceacf39bcfe02ff7aa289868c341366ee9f3a8e
289 lines
8.6 KiB
C++
289 lines
8.6 KiB
C++
// Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved.
|
|
// Copyright (c) 2015, Red Hat, Inc. All rights reserved.
|
|
// This source code is licensed under both the GPLv2 (found in the
|
|
// COPYING file in the root directory) and Apache 2.0 License
|
|
// (found in the LICENSE.Apache file in the root directory).
|
|
// Copyright (c) 2011 The LevelDB Authors. All rights reserved.
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file. See the AUTHORS file for names of contributors.
|
|
|
|
#include "rocksdb/utilities/env_mirror.h"
|
|
|
|
namespace ROCKSDB_NAMESPACE {
|
|
|
|
// An implementation of Env that mirrors all work over two backend
|
|
// Env's. This is useful for debugging purposes.
|
|
class SequentialFileMirror : public SequentialFile {
|
|
public:
|
|
std::unique_ptr<SequentialFile> a_, b_;
|
|
std::string fname;
|
|
explicit SequentialFileMirror(std::string f) : fname(f) {}
|
|
|
|
Status Read(size_t n, Slice* result, char* scratch) override {
|
|
Slice aslice;
|
|
Status as = a_->Read(n, &aslice, scratch);
|
|
if (as == Status::OK()) {
|
|
char* bscratch = new char[n];
|
|
Slice bslice;
|
|
#ifndef NDEBUG
|
|
size_t off = 0;
|
|
#endif
|
|
size_t left = aslice.size();
|
|
while (left) {
|
|
Status bs = b_->Read(left, &bslice, bscratch);
|
|
#ifndef NDEBUG
|
|
assert(as == bs);
|
|
assert(memcmp(bscratch, scratch + off, bslice.size()) == 0);
|
|
off += bslice.size();
|
|
#endif
|
|
left -= bslice.size();
|
|
}
|
|
delete[] bscratch;
|
|
*result = aslice;
|
|
} else {
|
|
Status bs = b_->Read(n, result, scratch);
|
|
assert(as == bs);
|
|
}
|
|
return as;
|
|
}
|
|
|
|
Status Skip(uint64_t n) override {
|
|
Status as = a_->Skip(n);
|
|
Status bs = b_->Skip(n);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status InvalidateCache(size_t offset, size_t length) override {
|
|
Status as = a_->InvalidateCache(offset, length);
|
|
Status bs = b_->InvalidateCache(offset, length);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
};
|
|
|
|
class RandomAccessFileMirror : public RandomAccessFile {
|
|
public:
|
|
std::unique_ptr<RandomAccessFile> a_, b_;
|
|
std::string fname;
|
|
explicit RandomAccessFileMirror(std::string f) : fname(f) {}
|
|
|
|
Status Read(uint64_t offset, size_t n, Slice* result,
|
|
char* scratch) const override {
|
|
Status as = a_->Read(offset, n, result, scratch);
|
|
if (as == Status::OK()) {
|
|
char* bscratch = new char[n];
|
|
Slice bslice;
|
|
size_t off = 0;
|
|
size_t left = result->size();
|
|
while (left) {
|
|
Status bs = b_->Read(offset + off, left, &bslice, bscratch);
|
|
assert(as == bs);
|
|
assert(memcmp(bscratch, scratch + off, bslice.size()) == 0);
|
|
off += bslice.size();
|
|
left -= bslice.size();
|
|
}
|
|
delete[] bscratch;
|
|
} else {
|
|
Status bs = b_->Read(offset, n, result, scratch);
|
|
assert(as == bs);
|
|
}
|
|
return as;
|
|
}
|
|
|
|
size_t GetUniqueId(char* id, size_t max_size) const override {
|
|
// NOTE: not verified
|
|
return a_->GetUniqueId(id, max_size);
|
|
}
|
|
|
|
Status GetFileSize(uint64_t* file_size) override {
|
|
uint64_t asize = 0, bsize = 0;
|
|
Status as = a_->GetFileSize(&asize);
|
|
Status bs = b_->GetFileSize(&bsize);
|
|
assert(as == bs);
|
|
assert(asize == bsize);
|
|
*file_size = asize;
|
|
return as;
|
|
}
|
|
};
|
|
|
|
class WritableFileMirror : public WritableFile {
|
|
public:
|
|
std::unique_ptr<WritableFile> a_, b_;
|
|
std::string fname;
|
|
explicit WritableFileMirror(std::string f, const EnvOptions& options)
|
|
: WritableFile(options), fname(f) {}
|
|
|
|
Status Append(const Slice& data) override {
|
|
Status as = a_->Append(data);
|
|
Status bs = b_->Append(data);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status Append(const Slice& data,
|
|
const DataVerificationInfo& /* verification_info */) override {
|
|
return Append(data);
|
|
}
|
|
Status PositionedAppend(const Slice& data, uint64_t offset) override {
|
|
Status as = a_->PositionedAppend(data, offset);
|
|
Status bs = b_->PositionedAppend(data, offset);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status PositionedAppend(
|
|
const Slice& data, uint64_t offset,
|
|
const DataVerificationInfo& /* verification_info */) override {
|
|
return PositionedAppend(data, offset);
|
|
}
|
|
Status Truncate(uint64_t size) override {
|
|
Status as = a_->Truncate(size);
|
|
Status bs = b_->Truncate(size);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status Close() override {
|
|
Status as = a_->Close();
|
|
Status bs = b_->Close();
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status Flush() override {
|
|
Status as = a_->Flush();
|
|
Status bs = b_->Flush();
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status Sync() override {
|
|
Status as = a_->Sync();
|
|
Status bs = b_->Sync();
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status Fsync() override {
|
|
Status as = a_->Fsync();
|
|
Status bs = b_->Fsync();
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
bool IsSyncThreadSafe() const override {
|
|
bool as = a_->IsSyncThreadSafe();
|
|
assert(as == b_->IsSyncThreadSafe());
|
|
return as;
|
|
}
|
|
void SetIOPriority(Env::IOPriority pri) override {
|
|
a_->SetIOPriority(pri);
|
|
b_->SetIOPriority(pri);
|
|
}
|
|
Env::IOPriority GetIOPriority() override {
|
|
// NOTE: we don't verify this one
|
|
return a_->GetIOPriority();
|
|
}
|
|
uint64_t GetFileSize() override {
|
|
uint64_t as = a_->GetFileSize();
|
|
assert(as == b_->GetFileSize());
|
|
return as;
|
|
}
|
|
void GetPreallocationStatus(size_t* block_size,
|
|
size_t* last_allocated_block) override {
|
|
// NOTE: we don't verify this one
|
|
return a_->GetPreallocationStatus(block_size, last_allocated_block);
|
|
}
|
|
size_t GetUniqueId(char* id, size_t max_size) const override {
|
|
// NOTE: we don't verify this one
|
|
return a_->GetUniqueId(id, max_size);
|
|
}
|
|
Status InvalidateCache(size_t offset, size_t length) override {
|
|
Status as = a_->InvalidateCache(offset, length);
|
|
Status bs = b_->InvalidateCache(offset, length);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
|
|
protected:
|
|
Status Allocate(uint64_t offset, uint64_t length) override {
|
|
Status as = a_->Allocate(offset, length);
|
|
Status bs = b_->Allocate(offset, length);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
Status RangeSync(uint64_t offset, uint64_t nbytes) override {
|
|
Status as = a_->RangeSync(offset, nbytes);
|
|
Status bs = b_->RangeSync(offset, nbytes);
|
|
assert(as == bs);
|
|
return as;
|
|
}
|
|
};
|
|
|
|
Status EnvMirror::NewSequentialFile(const std::string& f,
|
|
std::unique_ptr<SequentialFile>* r,
|
|
const EnvOptions& options) {
|
|
if (f.find("/proc/") == 0) {
|
|
return a_->NewSequentialFile(f, r, options);
|
|
}
|
|
SequentialFileMirror* mf = new SequentialFileMirror(f);
|
|
Status as = a_->NewSequentialFile(f, &mf->a_, options);
|
|
Status bs = b_->NewSequentialFile(f, &mf->b_, options);
|
|
assert(as == bs);
|
|
if (as.ok()) {
|
|
r->reset(mf);
|
|
} else {
|
|
delete mf;
|
|
}
|
|
return as;
|
|
}
|
|
|
|
Status EnvMirror::NewRandomAccessFile(const std::string& f,
|
|
std::unique_ptr<RandomAccessFile>* r,
|
|
const EnvOptions& options) {
|
|
if (f.find("/proc/") == 0) {
|
|
return a_->NewRandomAccessFile(f, r, options);
|
|
}
|
|
RandomAccessFileMirror* mf = new RandomAccessFileMirror(f);
|
|
Status as = a_->NewRandomAccessFile(f, &mf->a_, options);
|
|
Status bs = b_->NewRandomAccessFile(f, &mf->b_, options);
|
|
assert(as == bs);
|
|
if (as.ok()) {
|
|
r->reset(mf);
|
|
} else {
|
|
delete mf;
|
|
}
|
|
return as;
|
|
}
|
|
|
|
Status EnvMirror::NewWritableFile(const std::string& f,
|
|
std::unique_ptr<WritableFile>* r,
|
|
const EnvOptions& options) {
|
|
if (f.find("/proc/") == 0) {
|
|
return a_->NewWritableFile(f, r, options);
|
|
}
|
|
WritableFileMirror* mf = new WritableFileMirror(f, options);
|
|
Status as = a_->NewWritableFile(f, &mf->a_, options);
|
|
Status bs = b_->NewWritableFile(f, &mf->b_, options);
|
|
assert(as == bs);
|
|
if (as.ok()) {
|
|
r->reset(mf);
|
|
} else {
|
|
delete mf;
|
|
}
|
|
return as;
|
|
}
|
|
|
|
Status EnvMirror::ReuseWritableFile(const std::string& fname,
|
|
const std::string& old_fname,
|
|
std::unique_ptr<WritableFile>* r,
|
|
const EnvOptions& options) {
|
|
if (fname.find("/proc/") == 0) {
|
|
return a_->ReuseWritableFile(fname, old_fname, r, options);
|
|
}
|
|
WritableFileMirror* mf = new WritableFileMirror(fname, options);
|
|
Status as = a_->ReuseWritableFile(fname, old_fname, &mf->a_, options);
|
|
Status bs = b_->ReuseWritableFile(fname, old_fname, &mf->b_, options);
|
|
assert(as == bs);
|
|
if (as.ok()) {
|
|
r->reset(mf);
|
|
} else {
|
|
delete mf;
|
|
}
|
|
return as;
|
|
}
|
|
|
|
} // namespace ROCKSDB_NAMESPACE
|