93d155bc44
-- 3d20ce6cd6541579abecaba169d4b8716d511272 by Jon Cohen <cohenjon@google.com>: Only use LSAN for clang version >= 3.5. This should fix https://github.com/abseil/abseil-cpp/issues/244 PiperOrigin-RevId: 234675129 -- e15bd4ec7a81aa95cc3d09fa1e0e81d58ae478fb by Conrad Parker <conradparker@google.com>: Fix errors in apply() sample code The following changes are made: * Make the example method public. * Give the two user functions different names to avoid confusion about whether apply() can select the correct overload of a function based on its tuple argument (it can't). * Pass tuple2 to the second example apply() invocation, instead of passing its contents individually. * Fix a s/tuple/tuple3/ typo in the third example apply() invocation. PiperOrigin-RevId: 234223407 -- de0ed71e21bc76ddf9fe715fdbaef74cd0df95c7 by Abseil Team <absl-team@google.com>: First test if a macro is defined to avoid -Wundef. ABSL clients may need to compile their code with the -Wundef warning flag. It will be helpful if ABSL header files can be compiled without the -Wundef warning. How to avoid the -Wundef warning: If a macro may be undefined, we need to first test whether the macro is defined before testing its value. We can't rely on the C preprocessor rule that an undefined macro has the value 0L. PiperOrigin-RevId: 234201123 -- fa484ad7dae0cac21140a96662809ecb0ec8eb5d by Abseil Team <absl-team@google.com>: Internal change. PiperOrigin-RevId: 234185697 -- d69b1baef681e27954b065375ecf9c2320463b2b by Samuel Benzaquen <sbenza@google.com>: Mix pointers more thoroughly. Some pointer alignments interact badly with the mixing constant. By mixing twice we reduce this problem. PiperOrigin-RevId: 234178401 -- 1041d0e474610f3a8fea0db90958857327d6da1c by Samuel Benzaquen <sbenza@google.com>: Record rehashes in the hashtablez struct. Only recording the probe length on insertion causes a huge overestimation of the total probe length at any given time. With natural growth, elements are inserted when the load factor is between (max load/2, max load). However, after a rehash the majority of elements are actually inserted when the load factor is less than max/2 and have a much lower average probe length. Also reset some values when the table is cleared. PiperOrigin-RevId: 234013580 -- 299205caf3c89c47339f7409bc831746602cea84 by Mark Barolak <mbar@google.com>: Fix a sample code snippet that assumes `absl::string_view::const_iterator` is `const char*`. This is generally true, however in C++17 builds, absl::string_view is an alias for std::string_view and on MSVC, the std::string_view::const_iterator is an object instead of just a pointer. PiperOrigin-RevId: 233844595 -- af6c6370cf51a1e6c1469c79dfb2a486a4009136 by Abseil Team <absl-team@google.com>: Internal change. PiperOrigin-RevId: 233773470 -- 6e59e4b8e2bb6101b448f0f32b0bea81fe399ccf by Abseil Team <absl-team@google.com>: fix typo in {Starts|Ends}WithIgnoreCase comment in match.h PiperOrigin-RevId: 233662951 GitOrigin-RevId: 3d20ce6cd6541579abecaba169d4b8716d511272 Change-Id: Ib9a29b1c38c6aedf5d9f3f7f00596e8d30e864dd
355 lines
11 KiB
C++
355 lines
11 KiB
C++
// Copyright 2018 The Abseil Authors.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "absl/container/internal/hashtablez_sampler.h"
|
|
|
|
#include <atomic>
|
|
#include <limits>
|
|
#include <random>
|
|
|
|
#include "gmock/gmock.h"
|
|
#include "gtest/gtest.h"
|
|
#include "absl/base/attributes.h"
|
|
#include "absl/container/internal/have_sse.h"
|
|
#include "absl/synchronization/blocking_counter.h"
|
|
#include "absl/synchronization/internal/thread_pool.h"
|
|
#include "absl/synchronization/mutex.h"
|
|
#include "absl/synchronization/notification.h"
|
|
#include "absl/time/clock.h"
|
|
#include "absl/time/time.h"
|
|
|
|
#if SWISSTABLE_HAVE_SSE2
|
|
constexpr int kProbeLength = 16;
|
|
#else
|
|
constexpr int kProbeLength = 8;
|
|
#endif
|
|
|
|
namespace absl {
|
|
namespace container_internal {
|
|
class HashtablezInfoHandlePeer {
|
|
public:
|
|
static bool IsSampled(const HashtablezInfoHandle& h) {
|
|
return h.info_ != nullptr;
|
|
}
|
|
|
|
static HashtablezInfo* GetInfo(HashtablezInfoHandle* h) { return h->info_; }
|
|
};
|
|
|
|
namespace {
|
|
using ::absl::synchronization_internal::ThreadPool;
|
|
using ::testing::IsEmpty;
|
|
using ::testing::UnorderedElementsAre;
|
|
|
|
std::vector<size_t> GetSizes(HashtablezSampler* s) {
|
|
std::vector<size_t> res;
|
|
s->Iterate([&](const HashtablezInfo& info) {
|
|
res.push_back(info.size.load(std::memory_order_acquire));
|
|
});
|
|
return res;
|
|
}
|
|
|
|
HashtablezInfo* Register(HashtablezSampler* s, size_t size) {
|
|
auto* info = s->Register();
|
|
assert(info != nullptr);
|
|
info->size.store(size);
|
|
return info;
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, PrepareForSampling) {
|
|
absl::Time test_start = absl::Now();
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
|
|
EXPECT_EQ(info.capacity.load(), 0);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
EXPECT_EQ(info.max_probe_length.load(), 0);
|
|
EXPECT_EQ(info.total_probe_length.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), ~size_t{});
|
|
EXPECT_GE(info.create_time, test_start);
|
|
|
|
info.capacity.store(1, std::memory_order_relaxed);
|
|
info.size.store(1, std::memory_order_relaxed);
|
|
info.num_erases.store(1, std::memory_order_relaxed);
|
|
info.max_probe_length.store(1, std::memory_order_relaxed);
|
|
info.total_probe_length.store(1, std::memory_order_relaxed);
|
|
info.hashes_bitwise_or.store(1, std::memory_order_relaxed);
|
|
info.hashes_bitwise_and.store(1, std::memory_order_relaxed);
|
|
info.create_time = test_start - absl::Hours(20);
|
|
|
|
info.PrepareForSampling();
|
|
EXPECT_EQ(info.capacity.load(), 0);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
EXPECT_EQ(info.max_probe_length.load(), 0);
|
|
EXPECT_EQ(info.total_probe_length.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), ~size_t{});
|
|
EXPECT_GE(info.create_time, test_start);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordStorageChanged) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
RecordStorageChangedSlow(&info, 17, 47);
|
|
EXPECT_EQ(info.size.load(), 17);
|
|
EXPECT_EQ(info.capacity.load(), 47);
|
|
RecordStorageChangedSlow(&info, 20, 20);
|
|
EXPECT_EQ(info.size.load(), 20);
|
|
EXPECT_EQ(info.capacity.load(), 20);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordInsert) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
EXPECT_EQ(info.max_probe_length.load(), 0);
|
|
RecordInsertSlow(&info, 0x0000FF00, 6 * kProbeLength);
|
|
EXPECT_EQ(info.max_probe_length.load(), 6);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), 0x0000FF00);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0x0000FF00);
|
|
RecordInsertSlow(&info, 0x000FF000, 4 * kProbeLength);
|
|
EXPECT_EQ(info.max_probe_length.load(), 6);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), 0x0000F000);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0x000FFF00);
|
|
RecordInsertSlow(&info, 0x00FF0000, 12 * kProbeLength);
|
|
EXPECT_EQ(info.max_probe_length.load(), 12);
|
|
EXPECT_EQ(info.hashes_bitwise_and.load(), 0x00000000);
|
|
EXPECT_EQ(info.hashes_bitwise_or.load(), 0x00FFFF00);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordErase) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
RecordInsertSlow(&info, 0x0000FF00, 6 * kProbeLength);
|
|
EXPECT_EQ(info.size.load(), 1);
|
|
RecordEraseSlow(&info);
|
|
EXPECT_EQ(info.size.load(), 0);
|
|
EXPECT_EQ(info.num_erases.load(), 1);
|
|
}
|
|
|
|
TEST(HashtablezInfoTest, RecordRehash) {
|
|
HashtablezInfo info;
|
|
absl::MutexLock l(&info.init_mu);
|
|
info.PrepareForSampling();
|
|
RecordInsertSlow(&info, 0x1, 0);
|
|
RecordInsertSlow(&info, 0x2, kProbeLength);
|
|
RecordInsertSlow(&info, 0x4, kProbeLength);
|
|
RecordInsertSlow(&info, 0x8, 2 * kProbeLength);
|
|
EXPECT_EQ(info.size.load(), 4);
|
|
EXPECT_EQ(info.total_probe_length.load(), 4);
|
|
|
|
RecordEraseSlow(&info);
|
|
RecordEraseSlow(&info);
|
|
EXPECT_EQ(info.size.load(), 2);
|
|
EXPECT_EQ(info.total_probe_length.load(), 4);
|
|
EXPECT_EQ(info.num_erases.load(), 2);
|
|
|
|
RecordRehashSlow(&info, 3 * kProbeLength);
|
|
EXPECT_EQ(info.size.load(), 2);
|
|
EXPECT_EQ(info.total_probe_length.load(), 3);
|
|
EXPECT_EQ(info.num_erases.load(), 0);
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, SmallSampleParameter) {
|
|
SetHashtablezEnabled(true);
|
|
SetHashtablezSampleParameter(100);
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
|
int64_t next_sample = 0;
|
|
HashtablezInfo* sample = SampleSlow(&next_sample);
|
|
EXPECT_GT(next_sample, 0);
|
|
EXPECT_NE(sample, nullptr);
|
|
UnsampleSlow(sample);
|
|
}
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, LargeSampleParameter) {
|
|
SetHashtablezEnabled(true);
|
|
SetHashtablezSampleParameter(std::numeric_limits<int32_t>::max());
|
|
|
|
for (int i = 0; i < 1000; ++i) {
|
|
int64_t next_sample = 0;
|
|
HashtablezInfo* sample = SampleSlow(&next_sample);
|
|
EXPECT_GT(next_sample, 0);
|
|
EXPECT_NE(sample, nullptr);
|
|
UnsampleSlow(sample);
|
|
}
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Sample) {
|
|
SetHashtablezEnabled(true);
|
|
SetHashtablezSampleParameter(100);
|
|
int64_t num_sampled = 0;
|
|
int64_t total = 0;
|
|
double sample_rate;
|
|
for (int i = 0; i < 1000000; ++i) {
|
|
HashtablezInfoHandle h = Sample();
|
|
++total;
|
|
if (HashtablezInfoHandlePeer::IsSampled(h)) {
|
|
++num_sampled;
|
|
}
|
|
sample_rate = static_cast<double>(num_sampled) / total;
|
|
if (0.005 < sample_rate && sample_rate < 0.015) break;
|
|
}
|
|
EXPECT_NEAR(sample_rate, 0.01, 0.005);
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Handle) {
|
|
auto& sampler = HashtablezSampler::Global();
|
|
HashtablezInfoHandle h(sampler.Register());
|
|
auto* info = HashtablezInfoHandlePeer::GetInfo(&h);
|
|
info->hashes_bitwise_and.store(0x12345678, std::memory_order_relaxed);
|
|
|
|
bool found = false;
|
|
sampler.Iterate([&](const HashtablezInfo& h) {
|
|
if (&h == info) {
|
|
EXPECT_EQ(h.hashes_bitwise_and.load(), 0x12345678);
|
|
found = true;
|
|
}
|
|
});
|
|
EXPECT_TRUE(found);
|
|
|
|
h = HashtablezInfoHandle();
|
|
found = false;
|
|
sampler.Iterate([&](const HashtablezInfo& h) {
|
|
if (&h == info) {
|
|
// this will only happen if some other thread has resurrected the info
|
|
// the old handle was using.
|
|
if (h.hashes_bitwise_and.load() == 0x12345678) {
|
|
found = true;
|
|
}
|
|
}
|
|
});
|
|
EXPECT_FALSE(found);
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Registration) {
|
|
HashtablezSampler sampler;
|
|
auto* info1 = Register(&sampler, 1);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(1));
|
|
|
|
auto* info2 = Register(&sampler, 2);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(1, 2));
|
|
info1->size.store(3);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(3, 2));
|
|
|
|
sampler.Unregister(info1);
|
|
sampler.Unregister(info2);
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Unregistration) {
|
|
HashtablezSampler sampler;
|
|
std::vector<HashtablezInfo*> infos;
|
|
for (size_t i = 0; i < 3; ++i) {
|
|
infos.push_back(Register(&sampler, i));
|
|
}
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 1, 2));
|
|
|
|
sampler.Unregister(infos[1]);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 2));
|
|
|
|
infos.push_back(Register(&sampler, 3));
|
|
infos.push_back(Register(&sampler, 4));
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 2, 3, 4));
|
|
sampler.Unregister(infos[3]);
|
|
EXPECT_THAT(GetSizes(&sampler), UnorderedElementsAre(0, 2, 4));
|
|
|
|
sampler.Unregister(infos[0]);
|
|
sampler.Unregister(infos[2]);
|
|
sampler.Unregister(infos[4]);
|
|
EXPECT_THAT(GetSizes(&sampler), IsEmpty());
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, MultiThreaded) {
|
|
HashtablezSampler sampler;
|
|
Notification stop;
|
|
ThreadPool pool(10);
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
pool.Schedule([&sampler, &stop]() {
|
|
std::random_device rd;
|
|
std::mt19937 gen(rd());
|
|
|
|
std::vector<HashtablezInfo*> infoz;
|
|
while (!stop.HasBeenNotified()) {
|
|
if (infoz.empty()) {
|
|
infoz.push_back(sampler.Register());
|
|
}
|
|
switch (std::uniform_int_distribution<>(0, 2)(gen)) {
|
|
case 0: {
|
|
infoz.push_back(sampler.Register());
|
|
break;
|
|
}
|
|
case 1: {
|
|
size_t p =
|
|
std::uniform_int_distribution<>(0, infoz.size() - 1)(gen);
|
|
HashtablezInfo* info = infoz[p];
|
|
infoz[p] = infoz.back();
|
|
infoz.pop_back();
|
|
sampler.Unregister(info);
|
|
break;
|
|
}
|
|
case 2: {
|
|
absl::Duration oldest = absl::ZeroDuration();
|
|
sampler.Iterate([&](const HashtablezInfo& info) {
|
|
oldest = std::max(oldest, absl::Now() - info.create_time);
|
|
});
|
|
ASSERT_GE(oldest, absl::ZeroDuration());
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
});
|
|
}
|
|
// The threads will hammer away. Give it a little bit of time for tsan to
|
|
// spot errors.
|
|
absl::SleepFor(absl::Seconds(3));
|
|
stop.Notify();
|
|
}
|
|
|
|
TEST(HashtablezSamplerTest, Callback) {
|
|
HashtablezSampler sampler;
|
|
|
|
auto* info1 = Register(&sampler, 1);
|
|
auto* info2 = Register(&sampler, 2);
|
|
|
|
static const HashtablezInfo* expected;
|
|
|
|
auto callback = [](const HashtablezInfo& info) {
|
|
// We can't use `info` outside of this callback because the object will be
|
|
// disposed as soon as we return from here.
|
|
EXPECT_EQ(&info, expected);
|
|
};
|
|
|
|
// Set the callback.
|
|
EXPECT_EQ(sampler.SetDisposeCallback(callback), nullptr);
|
|
expected = info1;
|
|
sampler.Unregister(info1);
|
|
|
|
// Unset the callback.
|
|
EXPECT_EQ(callback, sampler.SetDisposeCallback(nullptr));
|
|
expected = nullptr; // no more calls.
|
|
sampler.Unregister(info2);
|
|
}
|
|
|
|
} // namespace
|
|
} // namespace container_internal
|
|
} // namespace absl
|