12bc53e031
-- c99f979ad34f155fbeeea69b88bdc7458d89a21c by Derek Mauro <dmauro@google.com>: Remove a floating point division by zero test. This isn't testing behavior related to the library, and MSVC warns about it in opt mode. PiperOrigin-RevId: 285220804 -- 68b015491f0dbf1ab547994673281abd1f34cd4b by Gennadiy Rozental <rogeeff@google.com>: This CL introduces following changes to the class FlagImpl: * We eliminate the CommandLineFlagLocks struct. Instead callback guard and callback function are combined into a single CallbackData struct, while primary data lock is stored separately. * CallbackData member of class FlagImpl is initially set to be nullptr and is only allocated and initialized when a flag's callback is being set. For most flags we do not pay for the extra space and extra absl::Mutex now. * Primary data guard is stored in data_guard_ data member. This is a properly aligned character buffer of necessary size. During initialization of the flag we construct absl::Mutex in this space using placement new call. * We now avoid extra value copy after successful attempt to parse value out of string. Instead we swap flag's current value with tentative value we just produced. PiperOrigin-RevId: 285132636 -- ed45d118fb818969eb13094cf7827c885dfc562c by Tom Manshreck <shreck@google.com>: Change null-term* (and nul-term*) to NUL-term* in comments PiperOrigin-RevId: 285036610 -- 729619017944db895ce8d6d29c1995aa2e5628a5 by Derek Mauro <dmauro@google.com>: Use the Posix implementation of thread identity on MinGW. Some versions of MinGW suffer from thread_local bugs. PiperOrigin-RevId: 285022920 -- 39a25493503c76885bc3254c28f66a251c5b5bb0 by Greg Falcon <gfalcon@google.com>: Implementation detail change. Add further ABSL_NAMESPACE_BEGIN and _END annotation macros to files in Abseil. PiperOrigin-RevId: 285012012 GitOrigin-RevId: c99f979ad34f155fbeeea69b88bdc7458d89a21c Change-Id: I4c85d3704e45d11a9ac50d562f39640a6adbedc1
359 lines
11 KiB
C++
359 lines
11 KiB
C++
// Copyright 2018 The Abseil Authors.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// https://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "absl/container/internal/hashtablez_sampler.h"
|
|
|
|
#include <atomic>
|
|
#include <limits>
|
|
#include <random>
|
|
|
|
#include "gmock/gmock.h"
|
|
#include "gtest/gtest.h"
|
|
#include "absl/base/attributes.h"
|
|
#include "absl/container/internal/have_sse.h"
|
|
#include "absl/synchronization/blocking_counter.h"
|
|
#include "absl/synchronization/internal/thread_pool.h"
|
|
#include "absl/synchronization/mutex.h"
|
|
#include "absl/synchronization/notification.h"
|
|
#include "absl/time/clock.h"
|
|
#include "absl/time/time.h"
|
|
|
|
#if SWISSTABLE_HAVE_SSE2
|
|
constexpr int kProbeLength = 16;
|
|
#else
|
|
constexpr int kProbeLength = 8;
|
|
#endif
|
|
|
|
namespace absl {
|
|
ABSL_NAMESPACE_BEGIN
|
|
namespace container_internal {
|
|
class HashtablezInfoHandlePeer {
|
|
public:
|
|
static bool IsSampled(const HashtablezInfoHandle& h) {
|
|
return h.info_ != nullptr;
|
|
}
|
|
|
|
static HashtablezInfo* GetInfo(HashtablezInfoHandle* h) { return h->info_; }
|
|
};
|
|
|
|
namespace {
|
|
using ::absl::synchronization_internal::ThreadPool;
|
|
using ::testing::IsEmpty;
|
|
using ::testing::UnorderedElementsAre;
|
|
|
|
std::vector<size_t> GetSizes(HashtablezSampler* s) {
|
|
std::vector<size_t> res;
|
|
s->Iterate([&](const HashtablezInfo& info) {
|
|
res.push_back(info.size.load(std::memory_order_acquire));
|
|
});
|
|
return res;
|
|
}
|
|
|
|
HashtablezInfo* Register(HashtablezSampler* s, size_t size) {
|
|
auto* info = s->Register();
|
|
assert(info != nullptr);
|
|
info->size.store(size);
|
|
return info;
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, PrepareForSampling) {
|
|
absl::Time test_start = absl::Now();
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
|
|
EXPECT_EQ(info.capacity.load(), 0);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
EXPECT_EQ(info.max_probe_length.load(), 0);
|
|
EXPECT_EQ(info.total_probe_length.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), ~size_t{});
|
|
EXPECT_GE(info.create_time, test_start);
|
|
|
|
info.capacity.store(1, std::memory_order_relaxed);
|
|
info.size.store(1, std::memory_order_relaxed);
|
|
info.num_erases.store(1, std::memory_order_relaxed);
|
|
info.max_probe_length.store(1, std::memory_order_relaxed);
|
|
info.total_probe_length.store(1, std::memory_order_relaxed);
|
|
info.hashes_bitwise_or.store(1, std::memory_order_relaxed);
|
|
info.hashes_bitwise_and.store(1, std::memory_order_relaxed);
|
|
info.create_time = test_start - absl::Hours(20);
|
|
|
|
info.PrepareForSampling();
|
|
EXPECT_EQ(info.capacity.load(), 0);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
EXPECT_EQ(info.max_probe_length.load(), 0);
|
|
EXPECT_EQ(info.total_probe_length.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), ~size_t{});
|
|
EXPECT_GE(info.create_time, test_start);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordStorageChanged) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
RecordStorageChangedSlow(&info, 17, 47);
|
|
EXPECT_EQ(info.size.load(), 17);
|
|
EXPECT_EQ(info.capacity.load(), 47);
|
|
RecordStorageChangedSlow(&info, 20, 20);
|
|
EXPECT_EQ(info.size.load(), 20);
|
|
EXPECT_EQ(info.capacity.load(), 20);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordInsert) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
EXPECT_EQ(info.max_probe_length.load(), 0);
|
|
RecordInsertSlow(&info, 0x0000FF00, 6 * kProbeLength);
|
|
EXPECT_EQ(info.max_probe_length.load(), 6);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), 0x0000FF00);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0x0000FF00);
|
|
RecordInsertSlow(&info, 0x000FF000, 4 * kProbeLength);
|
|
EXPECT_EQ(info.max_probe_length.load(), 6);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), 0x0000F000);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0x000FFF00);
|
|
RecordInsertSlow(&info, 0x00FF0000, 12 * kProbeLength);
|
|
EXPECT_EQ(info.max_probe_length.load(), 12);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), 0x00000000);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0x00FFFF00);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordErase) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
RecordInsertSlow(&info, 0x0000FF00, 6 * kProbeLength);
|
|
EXPECT_EQ(info.size.load(), 1);
|
|
RecordEraseSlow(&info);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
EXPECT_EQ(info.num_erases.load(), 1);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordRehash) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
RecordInsertSlow(&info, 0x1, 0);
|
|
RecordInsertSlow(&info, 0x2, kProbeLength);
|
|
RecordInsertSlow(&info, 0x4, kProbeLength);
|
|
RecordInsertSlow(&info, 0x8, 2 * kProbeLength);
|
|
EXPECT_EQ(info.size.load(), 4);
|
|
EXPECT_EQ(info.total_probe_length.load(), 4);
|
|
|
|
RecordEraseSlow(&info);
|
|
RecordEraseSlow(&info);
|
|
EXPECT_EQ(info.size.load(), 2);
|
|
EXPECT_EQ(info.total_probe_length.load(), 4);
|
|
EXPECT_EQ(info.num_erases.load(), 2);
|
|
|
|
RecordRehashSlow(&info, 3 * kProbeLength);
|
|
EXPECT_EQ(info.size.load(), 2);
|
|
EXPECT_EQ(info.total_probe_length.load(), 3);
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
}
|
|
|
|
#if ABSL_PER_THREAD_TLS == 1
|
|
TEST(HashtablezSamplerTest, SmallSampleParameter) {
|
|
SetHashtablezEnabled(true);
|
|
SetHashtablezSampleParameter(100);
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
|
int64_t next_sample = 0;
|
|
HashtablezInfo* sample = SampleSlow(&next_sample);
|
|
EXPECT_GT(next_sample, 0);
|
|
EXPECT_NE(sample, nullptr);
|
|
UnsampleSlow(sample);
|
|
}
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, LargeSampleParameter) {
|
|
SetHashtablezEnabled(true);
|
|
SetHashtablezSampleParameter(std::numeric_limits<int32_t>::max());
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
|
int64_t next_sample = 0;
|
|
HashtablezInfo* sample = SampleSlow(&next_sample);
|
|
EXPECT_GT(next_sample, 0);
|
|
EXPECT_NE(sample, nullptr);
|
|
UnsampleSlow(sample);
|
|
}
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Sample) {
|
|
SetHashtablezEnabled(true);
|
|
SetHashtablezSampleParameter(100);
|
|
int64_t num_sampled = 0;
|
|
int64_t total = 0;
|
|
double sample_rate = 0.0;
|
|
for (int i = 0; i < 1000000; ++i) {
|
|
HashtablezInfoHandle h = Sample();
|
|
++total;
|
|
if (HashtablezInfoHandlePeer::IsSampled(h)) {
|
|
++num_sampled;
|
|
}
|
|
sample_rate = static_cast<double>(num_sampled) / total;
|
|
if (0.005 < sample_rate && sample_rate < 0.015) break;
|
|
}
|
|
EXPECT_NEAR(sample_rate, 0.01, 0.005);
|
|
}
|
|
#endif
|
|
|
|
TEST(HashtablezSamplerTest, Handle) {
|
|
auto& sampler = HashtablezSampler::Global();
|
|
HashtablezInfoHandle h(sampler.Register());
|
|
auto* info = HashtablezInfoHandlePeer::GetInfo(&h);
|
|
info->hashes_bitwise_and.store(0x12345678, std::memory_order_relaxed);
|
|
|
|
bool found = false;
|
|
sampler.Iterate([&](const HashtablezInfo& h) {
|
|
if (&h == info) {
|
|
EXPECT_EQ(h.hashes_bitwise_and.load(), 0x12345678);
|
|
found = true;
|
|
}
|
|
});
|
|
EXPECT_TRUE(found);
|
|
|
|
h = HashtablezInfoHandle();
|
|
found = false;
|
|
sampler.Iterate([&](const HashtablezInfo& h) {
|
|
if (&h == info) {
|
|
// this will only happen if some other thread has resurrected the info
|
|
// the old handle was using.
|
|
if (h.hashes_bitwise_and.load() == 0x12345678) {
|
|
found = true;
|
|
}
|
|
}
|
|
});
|
|
EXPECT_FALSE(found);
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Registration) {
|
|
HashtablezSampler sampler;
|
|
auto* info1 = Register(&sampler, 1);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(1));
|
|
|
|
auto* info2 = Register(&sampler, 2);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(1, 2));
|
|
info1->size.store(3);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(3, 2));
|
|
|
|
sampler.Unregister(info1);
|
|
sampler.Unregister(info2);
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Unregistration) {
|
|
HashtablezSampler sampler;
|
|
std::vector<HashtablezInfo*> infos;
|
|
for (size_t i = 0; i < 3; ++i) {
|
|
infos.push_back(Register(&sampler, i));
|
|
}
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 1, 2));
|
|
|
|
sampler.Unregister(infos[1]);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 2));
|
|
|
|
infos.push_back(Register(&sampler, 3));
|
|
infos.push_back(Register(&sampler, 4));
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 2, 3, 4));
|
|
sampler.Unregister(infos[3]);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 2, 4));
|
|
|
|
sampler.Unregister(infos[0]);
|
|
sampler.Unregister(infos[2]);
|
|
sampler.Unregister(infos[4]);
|
|
EXPECT_THAT(GetSizes(&sampler), IsEmpty());
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, MultiThreaded) {
|
|
HashtablezSampler sampler;
|
|
Notification stop;
|
|
ThreadPool pool(10);
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
pool.Schedule([&sampler, &stop]() {
|
|
std::random_device rd;
|
|
std::mt19937 gen(rd());
|
|
|
|
std::vector<HashtablezInfo*> infoz;
|
|
while (!stop.HasBeenNotified()) {
|
|
if (infoz.empty()) {
|
|
infoz.push_back(sampler.Register());
|
|
}
|
|
switch (std::uniform_int_distribution<>(0, 2)(gen)) {
|
|
case 0: {
|
|
infoz.push_back(sampler.Register());
|
|
break;
|
|
}
|
|
case 1: {
|
|
size_t p =
|
|
std::uniform_int_distribution<>(0, infoz.size() - 1)(gen);
|
|
HashtablezInfo* info = infoz[p];
|
|
infoz[p] = infoz.back();
|
|
infoz.pop_back();
|
|
sampler.Unregister(info);
|
|
break;
|
|
}
|
|
case 2: {
|
|
absl::Duration oldest = absl::ZeroDuration();
|
|
sampler.Iterate([&](const HashtablezInfo& info) {
|
|
oldest = std::max(oldest, absl::Now() - info.create_time);
|
|
});
|
|
ASSERT_GE(oldest, absl::ZeroDuration());
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
});
|
|
}
|
|
// The threads will hammer away. Give it a little bit of time for tsan to
|
|
// spot errors.
|
|
absl::SleepFor(absl::Seconds(3));
|
|
stop.Notify();
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Callback) {
|
|
HashtablezSampler sampler;
|
|
|
|
auto* info1 = Register(&sampler, 1);
|
|
auto* info2 = Register(&sampler, 2);
|
|
|
|
static const HashtablezInfo* expected;
|
|
|
|
auto callback = [](const HashtablezInfo& info) {
|
|
// We can't use `info` outside of this callback because the object will be
|
|
// disposed as soon as we return from here.
|
|
EXPECT_EQ(&info, expected);
|
|
};
|
|
|
|
// Set the callback.
|
|
EXPECT_EQ(sampler.SetDisposeCallback(callback), nullptr);
|
|
expected = info1;
|
|
sampler.Unregister(info1);
|
|
|
|
// Unset the callback.
|
|
EXPECT_EQ(callback, sampler.SetDisposeCallback(nullptr));
|
|
expected = nullptr; // no more calls.
|
|
sampler.Unregister(info2);
|
|
}
|
|
|
|
} // namespace
|
|
} // namespace container_internal
|
|
ABSL_NAMESPACE_END
|
|
} // namespace absl
|