rocksdb/util/slice_test.cc
Peter Dillinger 4951494a27 Continue migration of HCC impl to BitFields (#14027)
Summary:
Continuing work from https://github.com/facebook/rocksdb/issues/13965. Here I'm migrating the "next with shift" kind of bit field and for that I've added an API for atomic additive transformations that can be combined into a single atomic update for multiple fields. (I implemented more features than needed, just in case they are needed someday and to demonstrate what is possible.)

Pull Request resolved: https://github.com/facebook/rocksdb/pull/14027

Test Plan: BitFields unit test updated/added, existing HCC tests

Reviewed By: xingbowang

Differential Revision: D83895094

Pulled By: pdillinger

fbshipit-source-id: e4487f34f5607b20f94b85a645ca654e6401e35d
2025-12-01 13:21:34 -08:00

659 lines
20 KiB
C++

// Copyright (c) 2011-present, Facebook, Inc. All rights reserved.
// This source code is licensed under both the GPLv2 (found in the
// COPYING file in the root directory) and Apache 2.0 License
// (found in the LICENSE.Apache file in the root directory).
// Because there are a small set of tests for Slice and there's a cost in having
// extra test binaries for each component, this test file has evolved into a
// "grab bag" of small tests for various reusable components, mostly in util/.
#include "rocksdb/slice.h"
#include <gtest/gtest.h>
#include <semaphore>
#include "port/port.h"
#include "port/stack_trace.h"
#include "rocksdb/data_structure.h"
#include "rocksdb/types.h"
#include "test_util/testharness.h"
#include "test_util/testutil.h"
#include "util/bit_fields.h"
#include "util/cast_util.h"
#include "util/semaphore.h"
#include "util/string_util.h"
namespace ROCKSDB_NAMESPACE {
TEST(SliceTest, StringView) {
std::string s = "foo";
std::string_view sv = s;
ASSERT_EQ(Slice(s), Slice(sv));
ASSERT_EQ(Slice(s), Slice(std::move(sv)));
}
// Use this to keep track of the cleanups that were actually performed
void Multiplier(void* arg1, void* arg2) {
int* res = static_cast<int*>(arg1);
int* num = static_cast<int*>(arg2);
*res *= *num;
}
class PinnableSliceTest : public testing::Test {
public:
void AssertSameData(const std::string& expected, const PinnableSlice& slice) {
std::string got;
got.assign(slice.data(), slice.size());
ASSERT_EQ(expected, got);
}
};
// Test that the external buffer is moved instead of being copied.
TEST_F(PinnableSliceTest, MoveExternalBuffer) {
Slice s("123");
std::string buf;
PinnableSlice v1(&buf);
v1.PinSelf(s);
PinnableSlice v2(std::move(v1));
ASSERT_EQ(buf.data(), v2.data());
ASSERT_EQ(&buf, v2.GetSelf());
PinnableSlice v3;
v3 = std::move(v2);
ASSERT_EQ(buf.data(), v3.data());
ASSERT_EQ(&buf, v3.GetSelf());
}
TEST_F(PinnableSliceTest, Move) {
int n2 = 2;
int res = 1;
const std::string const_str1 = "123";
const std::string const_str2 = "ABC";
Slice slice1(const_str1);
Slice slice2(const_str2);
{
// Test move constructor on a pinned slice.
res = 1;
PinnableSlice v1;
v1.PinSlice(slice1, Multiplier, &res, &n2);
PinnableSlice v2(std::move(v1));
// Since v1's Cleanable has been moved to v2,
// no cleanup should happen in Reset.
v1.Reset();
ASSERT_EQ(1, res);
AssertSameData(const_str1, v2);
}
// v2 is cleaned up.
ASSERT_EQ(2, res);
{
// Test move constructor on an unpinned slice.
PinnableSlice v1;
v1.PinSelf(slice1);
PinnableSlice v2(std::move(v1));
AssertSameData(const_str1, v2);
}
{
// Test move assignment from a pinned slice to
// another pinned slice.
res = 1;
PinnableSlice v1;
v1.PinSlice(slice1, Multiplier, &res, &n2);
PinnableSlice v2;
v2.PinSlice(slice2, Multiplier, &res, &n2);
v2 = std::move(v1);
// v2's Cleanable will be Reset before moving
// anything from v1.
ASSERT_EQ(2, res);
// Since v1's Cleanable has been moved to v2,
// no cleanup should happen in Reset.
v1.Reset();
ASSERT_EQ(2, res);
AssertSameData(const_str1, v2);
}
// The Cleanable moved from v1 to v2 will be Reset.
ASSERT_EQ(4, res);
{
// Test move assignment from a pinned slice to
// an unpinned slice.
res = 1;
PinnableSlice v1;
v1.PinSlice(slice1, Multiplier, &res, &n2);
PinnableSlice v2;
v2.PinSelf(slice2);
v2 = std::move(v1);
// Since v1's Cleanable has been moved to v2,
// no cleanup should happen in Reset.
v1.Reset();
ASSERT_EQ(1, res);
AssertSameData(const_str1, v2);
}
// The Cleanable moved from v1 to v2 will be Reset.
ASSERT_EQ(2, res);
{
// Test move assignment from an upinned slice to
// another unpinned slice.
PinnableSlice v1;
v1.PinSelf(slice1);
PinnableSlice v2;
v2.PinSelf(slice2);
v2 = std::move(v1);
AssertSameData(const_str1, v2);
}
{
// Test move assignment from an upinned slice to
// a pinned slice.
res = 1;
PinnableSlice v1;
v1.PinSelf(slice1);
PinnableSlice v2;
v2.PinSlice(slice2, Multiplier, &res, &n2);
v2 = std::move(v1);
// v2's Cleanable will be Reset before moving
// anything from v1.
ASSERT_EQ(2, res);
AssertSameData(const_str1, v2);
}
// No Cleanable is moved from v1 to v2, so no more cleanup.
ASSERT_EQ(2, res);
}
// ***************************************************************** //
// Unit test for SmallEnumSet
class SmallEnumSetTest : public testing::Test {
public:
SmallEnumSetTest() = default;
~SmallEnumSetTest() = default;
};
TEST_F(SmallEnumSetTest, SmallEnumSetTest1) {
FileTypeSet fs; // based on a legacy enum type
ASSERT_TRUE(fs.empty());
ASSERT_EQ(fs.count(), 0U);
ASSERT_TRUE(fs.Add(FileType::kIdentityFile));
ASSERT_FALSE(fs.empty());
ASSERT_EQ(fs.count(), 1U);
ASSERT_FALSE(fs.Add(FileType::kIdentityFile));
ASSERT_TRUE(fs.Add(FileType::kInfoLogFile));
ASSERT_TRUE(fs.Contains(FileType::kIdentityFile));
ASSERT_FALSE(fs.Contains(FileType::kDBLockFile));
ASSERT_FALSE(fs.empty());
ASSERT_EQ(fs.count(), 2U);
ASSERT_FALSE(fs.Remove(FileType::kDBLockFile));
ASSERT_TRUE(fs.Remove(FileType::kIdentityFile));
ASSERT_FALSE(fs.empty());
ASSERT_EQ(fs.count(), 1U);
ASSERT_TRUE(fs.Remove(FileType::kInfoLogFile));
ASSERT_TRUE(fs.empty());
ASSERT_EQ(fs.count(), 0U);
}
namespace {
enum class MyEnumClass { A, B, C };
} // namespace
using MyEnumClassSet = SmallEnumSet<MyEnumClass, MyEnumClass::C>;
TEST_F(SmallEnumSetTest, SmallEnumSetTest2) {
MyEnumClassSet s; // based on an enum class type
ASSERT_TRUE(s.Add(MyEnumClass::A));
ASSERT_TRUE(s.Contains(MyEnumClass::A));
ASSERT_FALSE(s.Contains(MyEnumClass::B));
ASSERT_TRUE(s.With(MyEnumClass::B).Contains(MyEnumClass::B));
ASSERT_TRUE(s.With(MyEnumClass::A).Contains(MyEnumClass::A));
ASSERT_FALSE(s.Contains(MyEnumClass::B));
ASSERT_FALSE(s.Without(MyEnumClass::A).Contains(MyEnumClass::A));
ASSERT_FALSE(
s.With(MyEnumClass::B).Without(MyEnumClass::B).Contains(MyEnumClass::B));
ASSERT_TRUE(
s.Without(MyEnumClass::B).With(MyEnumClass::B).Contains(MyEnumClass::B));
ASSERT_TRUE(s.Contains(MyEnumClass::A));
const MyEnumClassSet cs = s;
ASSERT_TRUE(cs.Contains(MyEnumClass::A));
ASSERT_EQ(cs, MyEnumClassSet{MyEnumClass::A});
ASSERT_EQ(cs.Without(MyEnumClass::A), MyEnumClassSet{});
ASSERT_EQ(cs, MyEnumClassSet::All().Without(MyEnumClass::B, MyEnumClass::C));
ASSERT_EQ(cs.With(MyEnumClass::B, MyEnumClass::C), MyEnumClassSet::All());
ASSERT_EQ(
MyEnumClassSet::All(),
MyEnumClassSet{}.With(MyEnumClass::A, MyEnumClass::B, MyEnumClass::C));
ASSERT_NE(cs, MyEnumClassSet{MyEnumClass::B});
ASSERT_NE(cs, MyEnumClassSet::All());
ASSERT_EQ(MyEnumClassSet{}.count(), 0U);
ASSERT_EQ(MyEnumClassSet::All().count(), 3U);
int count = 0;
for (MyEnumClass e : cs) {
ASSERT_EQ(e, MyEnumClass::A);
++count;
}
ASSERT_EQ(count, 1);
ASSERT_EQ(cs.count(), 1U);
count = 0;
for (MyEnumClass e : MyEnumClassSet::All().Without(MyEnumClass::B)) {
ASSERT_NE(e, MyEnumClass::B);
++count;
}
ASSERT_EQ(count, 2);
for (MyEnumClass e : MyEnumClassSet{}) {
(void)e;
assert(false);
}
}
template <typename ENUM_TYPE, ENUM_TYPE MAX_ENUMERATOR>
void TestBiggerEnumSet() {
using MySet = SmallEnumSet<ENUM_TYPE, MAX_ENUMERATOR>;
constexpr int kMaxValue = static_cast<int>(MAX_ENUMERATOR);
SCOPED_TRACE("kMaxValue = " + std::to_string(kMaxValue));
ASSERT_EQ(sizeof(MySet), (kMaxValue + 1 + 63) / 64 * 8);
MySet s;
ASSERT_TRUE(s.empty());
ASSERT_EQ(s.count(), 0U);
ASSERT_TRUE(s.Add(ENUM_TYPE(0)));
ASSERT_FALSE(s.empty());
ASSERT_EQ(s.count(), 1U);
ASSERT_TRUE(s.Add(ENUM_TYPE(kMaxValue - 1)));
ASSERT_FALSE(s.empty());
ASSERT_EQ(s.count(), 2U);
ASSERT_TRUE(s.Add(ENUM_TYPE(kMaxValue)));
ASSERT_FALSE(s.empty());
ASSERT_EQ(s.count(), 3U);
int count = 0;
for (ENUM_TYPE e : s) {
ASSERT_TRUE(e == ENUM_TYPE(0) || e == ENUM_TYPE(kMaxValue - 1) ||
e == ENUM_TYPE(kMaxValue));
++count;
}
ASSERT_EQ(count, 3);
ASSERT_TRUE(s.Remove(ENUM_TYPE(0)));
ASSERT_TRUE(s.Remove(ENUM_TYPE(kMaxValue)));
ASSERT_FALSE(s.empty());
ASSERT_EQ(s.count(), 1U);
count = 0;
for (ENUM_TYPE e : s) {
ASSERT_EQ(e, ENUM_TYPE(kMaxValue - 1));
++count;
}
ASSERT_EQ(count, 1);
}
TEST_F(SmallEnumSetTest, BiggerEnumClasses) {
enum class BiggerEnumClass63 { A, B, C = 63 };
enum class BiggerEnumClass64 { A, B, C = 64 };
enum class BiggerEnumClass65 { A, B, C = 65 };
enum class BiggerEnumClass127 { A, B, C = 127 };
enum class BiggerEnumClass128 { A, B, C = 128 };
enum class BiggerEnumClass129 { A, B, C = 129 };
enum class BiggerEnumClass150 { A, B, C = 150 };
enum class BiggerEnumClass255 { A, B, C = 255 };
TestBiggerEnumSet<BiggerEnumClass63, BiggerEnumClass63::C>();
TestBiggerEnumSet<BiggerEnumClass64, BiggerEnumClass64::C>();
TestBiggerEnumSet<BiggerEnumClass65, BiggerEnumClass65::C>();
TestBiggerEnumSet<BiggerEnumClass127, BiggerEnumClass127::C>();
TestBiggerEnumSet<BiggerEnumClass128, BiggerEnumClass128::C>();
TestBiggerEnumSet<BiggerEnumClass129, BiggerEnumClass129::C>();
TestBiggerEnumSet<BiggerEnumClass150, BiggerEnumClass150::C>();
TestBiggerEnumSet<BiggerEnumClass255, BiggerEnumClass255::C>();
}
// ***************************************************************** //
// Unit test for Status
TEST(StatusTest, Update) {
const Status ok = Status::OK();
const Status inc = Status::Incomplete("blah");
const Status notf = Status::NotFound("meow");
Status s = ok;
ASSERT_TRUE(s.UpdateIfOk(Status::Corruption("bad")).IsCorruption());
ASSERT_TRUE(s.IsCorruption());
s = ok;
ASSERT_TRUE(s.UpdateIfOk(Status::OK()).ok());
ASSERT_TRUE(s.UpdateIfOk(ok).ok());
ASSERT_TRUE(s.ok());
ASSERT_TRUE(s.UpdateIfOk(inc).IsIncomplete());
ASSERT_TRUE(s.IsIncomplete());
ASSERT_TRUE(s.UpdateIfOk(notf).IsIncomplete());
ASSERT_TRUE(s.UpdateIfOk(ok).IsIncomplete());
ASSERT_TRUE(s.IsIncomplete());
// Keeps left-most non-OK status
s = ok;
ASSERT_TRUE(
s.UpdateIfOk(Status()).UpdateIfOk(notf).UpdateIfOk(inc).IsNotFound());
ASSERT_TRUE(s.IsNotFound());
}
// ***************************************************************** //
// Unit test for UnownedPtr
TEST(UnownedPtrTest, Tests) {
{
int x = 0;
UnownedPtr<int> p(&x);
ASSERT_EQ(p.get(), &x);
ASSERT_EQ(*p, 0);
x = 1;
ASSERT_EQ(*p, 1);
ASSERT_EQ(p.get(), &x);
ASSERT_EQ(*p, 1);
*p = 2;
ASSERT_EQ(x, 2);
ASSERT_EQ(*p, 2);
ASSERT_EQ(p.get(), &x);
ASSERT_EQ(*p, 2);
}
{
std::unique_ptr<std::pair<int, int>> u =
std::make_unique<std::pair<int, int>>();
*u = {1, 2};
UnownedPtr<std::pair<int, int>> p;
ASSERT_FALSE(p);
p = u.get();
ASSERT_TRUE(p);
ASSERT_EQ(p->first, 1);
// These must not compile:
/*
u = p;
u = std::move(p);
std::unique_ptr<std::pair<int, int>> v{p};
std::unique_ptr<std::pair<int, int>> v{std::move(p)};
*/
// END must not compile
UnownedPtr<std::pair<int, int>> q;
q = std::move(p);
ASSERT_EQ(q->first, 1);
// Not committing to any moved-from semantics (on p here)
}
{
std::shared_ptr<std::pair<int, int>> s =
std::make_shared<std::pair<int, int>>();
*s = {1, 2};
UnownedPtr<std::pair<int, int>> p;
ASSERT_FALSE(p);
p = s.get();
ASSERT_TRUE(p);
ASSERT_EQ(p->first, 1);
// These must not compile:
/*
s = p;
s = std::move(p);
std::unique_ptr<std::pair<int, int>> t{p};
std::unique_ptr<std::pair<int, int>> t{std::move(p)};
*/
// END must not compile
UnownedPtr<std::pair<int, int>> q;
q = std::move(p);
ASSERT_EQ(q->first, 1);
// Not committing to any moved-from semantics (on p here)
}
}
TEST(ToBaseCharsStringTest, Tests) {
using ROCKSDB_NAMESPACE::ToBaseCharsString;
// Base 16
ASSERT_EQ(ToBaseCharsString<16>(5, 0, true), "00000");
ASSERT_EQ(ToBaseCharsString<16>(5, 42, true), "0002A");
ASSERT_EQ(ToBaseCharsString<16>(5, 42, false), "0002a");
ASSERT_EQ(ToBaseCharsString<16>(2, 255, false), "ff");
// Base 32
ASSERT_EQ(ToBaseCharsString<32>(2, 255, false), "7v");
}
TEST(SemaphoreTest, CountingSemaphore) {
CountingSemaphore sem{0};
int kCount = 5;
std::vector<std::thread> threads;
for (int i = 0; i < kCount; ++i) {
threads.emplace_back([&sem] { sem.Release(); });
}
for (int i = 0; i < kCount; ++i) {
threads.emplace_back([&sem] { sem.Acquire(); });
}
for (auto& t : threads) {
t.join();
}
// Nothing left on the semaphore
ASSERT_FALSE(sem.TryAcquire());
// Keep testing
sem.Release(2);
ASSERT_TRUE(sem.TryAcquire());
sem.Acquire();
ASSERT_FALSE(sem.TryAcquire());
}
TEST(SemaphoreTest, BinarySemaphore) {
BinarySemaphore sem{0};
int kCount = 5;
std::vector<std::thread> threads;
for (int i = 0; i < kCount; ++i) {
threads.emplace_back([&sem] {
sem.Acquire();
sem.Release();
});
}
threads.emplace_back([&sem] { sem.Release(); });
for (auto& t : threads) {
t.join();
}
// Only able to acquire one excess release
ASSERT_TRUE(sem.TryAcquire());
ASSERT_FALSE(sem.TryAcquire());
}
TEST(BitFieldsTest, BitFields) {
// Start by verifying example from BitFields comment
struct MyState : public BitFields<uint32_t, MyState> {
// Extra helper declarations and/or field type declarations
};
using Field1 = UnsignedBitField<MyState, 16, NoPrevBitField>;
using Field2 = BoolBitField<MyState, Field1>;
using Field3 = BoolBitField<MyState, Field2>;
using Field4 = UnsignedBitField<MyState, 5, Field3>;
// MyState{} is zero-initialized
auto state = MyState{}.With<Field1>(42U).With<Field2>(true);
state.Set<Field4>(3U);
state.Ref<Field1>() += state.Get<Field4>();
ASSERT_EQ(state.Get<Field1>(), 45U);
ASSERT_EQ(state.Get<Field2>(), true);
ASSERT_EQ(state.Get<Field3>(), false);
ASSERT_EQ(state.Get<Field4>(), 3U);
// Misc operators
auto ref = state.Ref<Field3>();
auto ref2 = std::move(ref);
ref2 = true;
ASSERT_EQ(state.Get<Field3>(), true);
MyState state2;
// Basic non-concurrent tests for atomic wrappers
{
RelaxedBitFieldsAtomic<MyState> relaxed{state};
ASSERT_EQ(state, relaxed.LoadRelaxed());
relaxed.StoreRelaxed(state2);
ASSERT_EQ(state2, relaxed.LoadRelaxed());
MyState state3 = relaxed.ExchangeRelaxed(state);
ASSERT_EQ(state2, state3);
ASSERT_TRUE(relaxed.CasStrongRelaxed(state, state2));
while (!relaxed.CasWeakRelaxed(state2, state)) {
}
ASSERT_EQ(state2, state3);
ASSERT_EQ(state, relaxed.LoadRelaxed());
auto transform1 = Field2::ClearTransform() + Field3::ClearTransform();
MyState before, after;
relaxed.ApplyRelaxed(transform1, &before, &after);
ASSERT_EQ(before, state);
ASSERT_NE(after, state);
ASSERT_EQ(after.Get<Field2>(), false);
ASSERT_EQ(after.Get<Field3>(), false);
auto transform2 = Field2::SetTransform() + Field3::SetTransform();
relaxed.ApplyRelaxed(transform2, &before, &after);
ASSERT_NE(before, state);
ASSERT_EQ(before.Get<Field2>(), false);
ASSERT_EQ(before.Get<Field3>(), false);
ASSERT_EQ(after, state);
ASSERT_EQ(state.Get<Field1>(), 45U);
ASSERT_EQ(after.Get<Field2>(), true);
ASSERT_EQ(after.Get<Field3>(), true);
ASSERT_EQ(state.Get<Field4>(), 3U);
auto transform3 = Field1::PlusTransformPromiseNoOverflow(10000U) +
Field4::MinusTransformPromiseNoUnderflow(3U);
relaxed.ApplyRelaxed(transform3, &before, &after);
ASSERT_EQ(before, state);
ASSERT_NE(after, state);
ASSERT_EQ(after.Get<Field1>(), 10045U);
ASSERT_EQ(after.Get<Field4>(), 0U);
auto transform4 = Field1::MinusTransformPromiseNoUnderflow(999U) +
Field4::PlusTransformPromiseNoOverflow(31U);
relaxed.ApplyRelaxed(transform4, &before, &after);
ASSERT_EQ(after.Get<Field1>(), 9046U);
ASSERT_EQ(after.Get<Field4>(), 31U);
// Unmodified
ASSERT_EQ(after.Get<Field2>(), true);
ASSERT_EQ(after.Get<Field3>(), true);
// Test overflow/underflow detection
relaxed.StoreRelaxed(MyState{}.With<Field1>(65535U)); // Field1 max value
ASSERT_TESTABLE_FAILURE(
relaxed.ApplyRelaxed(Field1::PlusTransformPromiseNoOverflow(1U)));
relaxed.StoreRelaxed(MyState{}.With<Field4>(31U)); // Field4 max value
ASSERT_TESTABLE_FAILURE(
relaxed.ApplyRelaxed(Field4::PlusTransformPromiseNoOverflow(1U)));
relaxed.StoreRelaxed(MyState{}.With<Field1>(0U));
ASSERT_TESTABLE_FAILURE(
relaxed.ApplyRelaxed(Field1::MinusTransformPromiseNoUnderflow(1U)));
relaxed.StoreRelaxed(MyState{}.With<Field4>(0U));
ASSERT_TESTABLE_FAILURE(
relaxed.ApplyRelaxed(Field4::MinusTransformPromiseNoUnderflow(1U)));
ASSERT_TESTABLE_FAILURE(relaxed.ApplyRelaxed(
Field4::MinusTransformPromiseNoUnderflow(64U))); // Too big
ASSERT_TESTABLE_FAILURE(relaxed.ApplyRelaxed(
Field4::PlusTransformPromiseNoOverflow(64U))); // Too big
// Including combinations
relaxed.StoreRelaxed(MyState{}.With<Field4>(31U)); // Field4 max value
relaxed.StoreRelaxed(MyState{}.With<Field1>(0U));
ASSERT_TESTABLE_FAILURE(
relaxed.ApplyRelaxed(Field4::PlusTransformPromiseNoOverflow(1U) +
Field1::MinusTransformPromiseNoUnderflow(1U)));
// But a field at the limit of upper bits is allowed to over/underflow
using Field5 = UnsignedBitField<MyState, 9, Field4>;
relaxed.StoreRelaxed(MyState{}.With<Field5>(0)); // Field5 max value
relaxed.ApplyRelaxed(Field5::MinusTransformIgnoreUnderflow(1U), &before,
&after); // "Safe" underflow
ASSERT_EQ(after.Get<Field5>(), 511U);
relaxed.ApplyRelaxed(Field5::PlusTransformIgnoreOverflow(1U), &before,
&after); // "Safe" overflow
ASSERT_EQ(after.Get<Field5>(), 0U);
relaxed.ApplyRelaxed(Field5::PlusTransformIgnoreOverflow(2048U), &before,
&after); // "Safe" overflow
ASSERT_EQ(after.Get<Field5>(), 0U);
}
{
AcqRelBitFieldsAtomic<MyState> acqrel{state};
ASSERT_EQ(state, acqrel.Load());
acqrel.Store(state2);
ASSERT_EQ(state2, acqrel.Load());
MyState state3 = acqrel.Exchange(state);
ASSERT_EQ(state2, state3);
ASSERT_TRUE(acqrel.CasStrong(state, state2));
while (!acqrel.CasWeak(state2, state)) {
}
ASSERT_EQ(state2, state3);
ASSERT_EQ(state, acqrel.Load());
auto transform1 = Field2::ClearTransform() + Field3::ClearTransform();
MyState before, after;
acqrel.Apply(transform1, &before, &after);
ASSERT_EQ(before, state);
ASSERT_NE(after, state);
ASSERT_EQ(after.Get<Field2>(), false);
ASSERT_EQ(after.Get<Field3>(), false);
auto transform2 = Field2::SetTransform() + Field3::SetTransform();
acqrel.Apply(transform2, &before, &after);
ASSERT_NE(before, state);
ASSERT_EQ(before.Get<Field2>(), false);
ASSERT_EQ(before.Get<Field3>(), false);
ASSERT_EQ(after, state);
ASSERT_EQ(state.Get<Field1>(), 45U);
ASSERT_EQ(after.Get<Field2>(), true);
ASSERT_EQ(after.Get<Field3>(), true);
ASSERT_EQ(state.Get<Field4>(), 3U);
auto transform3 = Field1::PlusTransformPromiseNoOverflow(10000U) +
Field4::MinusTransformPromiseNoUnderflow(3U);
acqrel.Apply(transform3, &before, &after);
ASSERT_EQ(before, state);
ASSERT_NE(after, state);
ASSERT_EQ(after.Get<Field1>(), 10045U);
ASSERT_EQ(after.Get<Field4>(), 0U);
auto transform4 = Field1::MinusTransformPromiseNoUnderflow(999U) +
Field4::PlusTransformPromiseNoOverflow(31U);
acqrel.Apply(transform4, &before, &after);
ASSERT_EQ(after.Get<Field1>(), 9046U);
ASSERT_EQ(after.Get<Field4>(), 31U);
// Unmodified
ASSERT_EQ(after.Get<Field2>(), true);
ASSERT_EQ(after.Get<Field3>(), true);
// Test overflow/underflow detection
acqrel.Store(MyState{}.With<Field1>(65535U));
ASSERT_TESTABLE_FAILURE(
acqrel.Apply(Field1::PlusTransformPromiseNoOverflow(1U)));
acqrel.Store(MyState{}.With<Field4>(0U));
ASSERT_TESTABLE_FAILURE(
acqrel.Apply(Field4::MinusTransformPromiseNoUnderflow(1U)));
}
}
} // namespace ROCKSDB_NAMESPACE
int main(int argc, char** argv) {
ROCKSDB_NAMESPACE::port::InstallStackTraceHandler();
::testing::InitGoogleTest(&argc, argv);
return RUN_ALL_TESTS();
}