ca9856cabc
-- 53550735f5a943dfb99225e7c53f211c2d6e7951 by Gennadiy Rozental <rogeeff@google.com>: Import of CCTZ from GitHub. PiperOrigin-RevId: 309333648 -- 847bbf8a1d9cd322ec058c6f932d1f687fd3d331 by Gennadiy Rozental <rogeeff@google.com>: Make Validation interfaces private in CommandLineFlag. Calls are rewired via private interface access struct. PiperOrigin-RevId: 309323013 -- a600fc5051e0a0af50a7850450fd3ed1aef3f316 by Matthew Brown <matthewbr@google.com>: Internal Change. PiperOrigin-RevId: 309292207 -- 937d00ce3cf62c5f23f59b5377471fd01d6bfbc7 by Gennadiy Rozental <rogeeff@google.com>: Make TypeId interface private in CommandLineFlag. We also rewire the SaveState via the new PrivateHandleInterface trampoline class. This class will be the only way to access private methods of class CommandLineFlag. PiperOrigin-RevId: 309282547 -- 796c4bd35073b6a8337762bdb13603dae12a4df1 by Derek Mauro <dmauro@google.com>: Cleanup uses of kLinkerInitialized PiperOrigin-RevId: 309274734 -- c831446c52d9ef4bdcb1ea369840904620abc4b9 by Gennadiy Rozental <rogeeff@google.com>: Eliminate the interface IsModified of CommndLineFlag. PiperOrigin-RevId: 309256248 -- a1db59d7f7aa39cb0a37dbf80f8c04e371da8465 by Gennadiy Rozental <rogeeff@google.com>: Avoid default value generator if default value expression is constexpr. If possible, we detect constexpr-ness of default value expression and avoid storing default value generator in side of flag and instead set the flag's value to the value of that expression at const initialization time of flag objects. At the moment we only do this for flags of (all) integral, float and double value types PiperOrigin-RevId: 309110630 -- ae3b4a139aacd8fc165c9acd2a3cbae1f9e26af4 by Gennadiy Rozental <rogeeff@google.com>: Make SaveState a private method of the CommandLineFlag and make it only accessible from FlagSaverImpl. There is no other call sites for this call. PiperOrigin-RevId: 309073989 -- cbc24b4dcc166dd6b0208e9d7620484eaaaa7ee0 by Abseil Team <absl-team@google.com>: Eliminate the interface IsModified of CommndLineFlag. PiperOrigin-RevId: 309064639 -- 08e79645a89d71785c5381cea9c413357db9824a by Gennadiy Rozental <rogeeff@google.com>: Eliminate the interface IsModified of CommndLineFlag. PiperOrigin-RevId: 309054430 -- 4a6c70233c60dc8c39b7fa9beb5fa687c215261f by Gennadiy Rozental <rogeeff@google.com>: Internal change PiperOrigin-RevId: 308900784 -- 13160efdf7710f142778d5a1e4c85aa309f019b6 by Abseil Team <absl-team@google.com>: Provide definitions of static member variables -- improved C++11 support. PiperOrigin-RevId: 308900290 -- 0343b8228657b9b313afdfe88c4a7b2137d56db4 by Gennadiy Rozental <rogeeff@google.com>: Rename method Get<T> to TryGet<T> per approved spec before making interface public. PiperOrigin-RevId: 308889113 -- 7b84e27fb857fc1296a05504970f506d47d2f2c1 by Derek Mauro <dmauro@google.com>: Remove node_hash_* methods that were deprecated on release PiperOrigin-RevId: 308837933 -- 599d44ee72c02b6bb6e1c1a1db72873841441416 by Gennadiy Rozental <rogeeff@google.com>: Eliminate CommandLineFlag::Typename interface per approved spec before making CommandLineFlag public. PiperOrigin-RevId: 308814376 GitOrigin-RevId: 53550735f5a943dfb99225e7c53f211c2d6e7951 Change-Id: Iae52c65b7322152c7e58f222d60eb5a21699a2cb
265 lines
9.3 KiB
C++
265 lines
9.3 KiB
C++
// Copyright 2017 The Abseil Authors.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// https://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
// A bunch of threads repeatedly hash an array of ints protected by a
|
|
// spinlock. If the spinlock is working properly, all elements of the
|
|
// array should be equal at the end of the test.
|
|
|
|
#include <cstdint>
|
|
#include <limits>
|
|
#include <random>
|
|
#include <thread> // NOLINT(build/c++11)
|
|
#include <vector>
|
|
|
|
#include "gtest/gtest.h"
|
|
#include "absl/base/attributes.h"
|
|
#include "absl/base/internal/low_level_scheduling.h"
|
|
#include "absl/base/internal/scheduling_mode.h"
|
|
#include "absl/base/internal/spinlock.h"
|
|
#include "absl/base/internal/sysinfo.h"
|
|
#include "absl/base/macros.h"
|
|
#include "absl/synchronization/blocking_counter.h"
|
|
#include "absl/synchronization/notification.h"
|
|
|
|
constexpr int32_t kNumThreads = 10;
|
|
constexpr int32_t kIters = 1000;
|
|
|
|
namespace absl {
|
|
ABSL_NAMESPACE_BEGIN
|
|
namespace base_internal {
|
|
|
|
// This is defined outside of anonymous namespace so that it can be
|
|
// a friend of SpinLock to access protected methods for testing.
|
|
struct SpinLockTest {
|
|
static uint32_t EncodeWaitCycles(int64_t wait_start_time,
|
|
int64_t wait_end_time) {
|
|
return SpinLock::EncodeWaitCycles(wait_start_time, wait_end_time);
|
|
}
|
|
static uint64_t DecodeWaitCycles(uint32_t lock_value) {
|
|
return SpinLock::DecodeWaitCycles(lock_value);
|
|
}
|
|
};
|
|
|
|
namespace {
|
|
|
|
static constexpr int kArrayLength = 10;
|
|
static uint32_t values[kArrayLength];
|
|
|
|
ABSL_CONST_INIT static SpinLock static_cooperative_spinlock(
|
|
absl::kConstInit, base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL);
|
|
ABSL_CONST_INIT static SpinLock static_noncooperative_spinlock(
|
|
absl::kConstInit, base_internal::SCHEDULE_KERNEL_ONLY);
|
|
|
|
// Simple integer hash function based on the public domain lookup2 hash.
|
|
// http://burtleburtle.net/bob/c/lookup2.c
|
|
static uint32_t Hash32(uint32_t a, uint32_t c) {
|
|
uint32_t b = 0x9e3779b9UL; // The golden ratio; an arbitrary value.
|
|
a -= b; a -= c; a ^= (c >> 13);
|
|
b -= c; b -= a; b ^= (a << 8);
|
|
c -= a; c -= b; c ^= (b >> 13);
|
|
a -= b; a -= c; a ^= (c >> 12);
|
|
b -= c; b -= a; b ^= (a << 16);
|
|
c -= a; c -= b; c ^= (b >> 5);
|
|
a -= b; a -= c; a ^= (c >> 3);
|
|
b -= c; b -= a; b ^= (a << 10);
|
|
c -= a; c -= b; c ^= (b >> 15);
|
|
return c;
|
|
}
|
|
|
|
static void TestFunction(int thread_salt, SpinLock* spinlock) {
|
|
for (int i = 0; i < kIters; i++) {
|
|
SpinLockHolder h(spinlock);
|
|
for (int j = 0; j < kArrayLength; j++) {
|
|
const int index = (j + thread_salt) % kArrayLength;
|
|
values[index] = Hash32(values[index], thread_salt);
|
|
std::this_thread::yield();
|
|
}
|
|
}
|
|
}
|
|
|
|
static void ThreadedTest(SpinLock* spinlock) {
|
|
std::vector<std::thread> threads;
|
|
for (int i = 0; i < kNumThreads; ++i) {
|
|
threads.push_back(std::thread(TestFunction, i, spinlock));
|
|
}
|
|
for (auto& thread : threads) {
|
|
thread.join();
|
|
}
|
|
|
|
SpinLockHolder h(spinlock);
|
|
for (int i = 1; i < kArrayLength; i++) {
|
|
EXPECT_EQ(values[0], values[i]);
|
|
}
|
|
}
|
|
|
|
TEST(SpinLock, StackNonCooperativeDisablesScheduling) {
|
|
SpinLock spinlock(base_internal::SCHEDULE_KERNEL_ONLY);
|
|
spinlock.Lock();
|
|
EXPECT_FALSE(base_internal::SchedulingGuard::ReschedulingIsAllowed());
|
|
spinlock.Unlock();
|
|
}
|
|
|
|
TEST(SpinLock, StaticNonCooperativeDisablesScheduling) {
|
|
static_noncooperative_spinlock.Lock();
|
|
EXPECT_FALSE(base_internal::SchedulingGuard::ReschedulingIsAllowed());
|
|
static_noncooperative_spinlock.Unlock();
|
|
}
|
|
|
|
TEST(SpinLock, WaitCyclesEncoding) {
|
|
// These are implementation details not exported by SpinLock.
|
|
const int kProfileTimestampShift = 7;
|
|
const int kLockwordReservedShift = 3;
|
|
const uint32_t kSpinLockSleeper = 8;
|
|
|
|
// We should be able to encode up to (1^kMaxCycleBits - 1) without clamping
|
|
// but the lower kProfileTimestampShift will be dropped.
|
|
const int kMaxCyclesShift =
|
|
32 - kLockwordReservedShift + kProfileTimestampShift;
|
|
const uint64_t kMaxCycles = (int64_t{1} << kMaxCyclesShift) - 1;
|
|
|
|
// These bits should be zero after encoding.
|
|
const uint32_t kLockwordReservedMask = (1 << kLockwordReservedShift) - 1;
|
|
|
|
// These bits are dropped when wait cycles are encoded.
|
|
const uint64_t kProfileTimestampMask = (1 << kProfileTimestampShift) - 1;
|
|
|
|
// Test a bunch of random values
|
|
std::default_random_engine generator;
|
|
// Shift to avoid overflow below.
|
|
std::uniform_int_distribution<uint64_t> time_distribution(
|
|
0, std::numeric_limits<uint64_t>::max() >> 4);
|
|
std::uniform_int_distribution<uint64_t> cycle_distribution(0, kMaxCycles);
|
|
|
|
for (int i = 0; i < 100; i++) {
|
|
int64_t start_time = time_distribution(generator);
|
|
int64_t cycles = cycle_distribution(generator);
|
|
int64_t end_time = start_time + cycles;
|
|
uint32_t lock_value = SpinLockTest::EncodeWaitCycles(start_time, end_time);
|
|
EXPECT_EQ(0, lock_value & kLockwordReservedMask);
|
|
uint64_t decoded = SpinLockTest::DecodeWaitCycles(lock_value);
|
|
EXPECT_EQ(0, decoded & kProfileTimestampMask);
|
|
EXPECT_EQ(cycles & ~kProfileTimestampMask, decoded);
|
|
}
|
|
|
|
// Test corner cases
|
|
int64_t start_time = time_distribution(generator);
|
|
EXPECT_EQ(kSpinLockSleeper,
|
|
SpinLockTest::EncodeWaitCycles(start_time, start_time));
|
|
EXPECT_EQ(0, SpinLockTest::DecodeWaitCycles(0));
|
|
EXPECT_EQ(0, SpinLockTest::DecodeWaitCycles(kLockwordReservedMask));
|
|
EXPECT_EQ(kMaxCycles & ~kProfileTimestampMask,
|
|
SpinLockTest::DecodeWaitCycles(~kLockwordReservedMask));
|
|
|
|
// Check that we cannot produce kSpinLockSleeper during encoding.
|
|
int64_t sleeper_cycles =
|
|
kSpinLockSleeper << (kProfileTimestampShift - kLockwordReservedShift);
|
|
uint32_t sleeper_value =
|
|
SpinLockTest::EncodeWaitCycles(start_time, start_time + sleeper_cycles);
|
|
EXPECT_NE(sleeper_value, kSpinLockSleeper);
|
|
|
|
// Test clamping
|
|
uint32_t max_value =
|
|
SpinLockTest::EncodeWaitCycles(start_time, start_time + kMaxCycles);
|
|
uint64_t max_value_decoded = SpinLockTest::DecodeWaitCycles(max_value);
|
|
uint64_t expected_max_value_decoded = kMaxCycles & ~kProfileTimestampMask;
|
|
EXPECT_EQ(expected_max_value_decoded, max_value_decoded);
|
|
|
|
const int64_t step = (1 << kProfileTimestampShift);
|
|
uint32_t after_max_value =
|
|
SpinLockTest::EncodeWaitCycles(start_time, start_time + kMaxCycles + step);
|
|
uint64_t after_max_value_decoded =
|
|
SpinLockTest::DecodeWaitCycles(after_max_value);
|
|
EXPECT_EQ(expected_max_value_decoded, after_max_value_decoded);
|
|
|
|
uint32_t before_max_value = SpinLockTest::EncodeWaitCycles(
|
|
start_time, start_time + kMaxCycles - step);
|
|
uint64_t before_max_value_decoded =
|
|
SpinLockTest::DecodeWaitCycles(before_max_value);
|
|
EXPECT_GT(expected_max_value_decoded, before_max_value_decoded);
|
|
}
|
|
|
|
TEST(SpinLockWithThreads, StackSpinLock) {
|
|
SpinLock spinlock;
|
|
ThreadedTest(&spinlock);
|
|
}
|
|
|
|
TEST(SpinLockWithThreads, StackCooperativeSpinLock) {
|
|
SpinLock spinlock(base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL);
|
|
ThreadedTest(&spinlock);
|
|
}
|
|
|
|
TEST(SpinLockWithThreads, StackNonCooperativeSpinLock) {
|
|
SpinLock spinlock(base_internal::SCHEDULE_KERNEL_ONLY);
|
|
ThreadedTest(&spinlock);
|
|
}
|
|
|
|
TEST(SpinLockWithThreads, StaticCooperativeSpinLock) {
|
|
ThreadedTest(&static_cooperative_spinlock);
|
|
}
|
|
|
|
TEST(SpinLockWithThreads, StaticNonCooperativeSpinLock) {
|
|
ThreadedTest(&static_noncooperative_spinlock);
|
|
}
|
|
|
|
TEST(SpinLockWithThreads, DoesNotDeadlock) {
|
|
struct Helper {
|
|
static void NotifyThenLock(Notification* locked, SpinLock* spinlock,
|
|
BlockingCounter* b) {
|
|
locked->WaitForNotification(); // Wait for LockThenWait() to hold "s".
|
|
b->DecrementCount();
|
|
SpinLockHolder l(spinlock);
|
|
}
|
|
|
|
static void LockThenWait(Notification* locked, SpinLock* spinlock,
|
|
BlockingCounter* b) {
|
|
SpinLockHolder l(spinlock);
|
|
locked->Notify();
|
|
b->Wait();
|
|
}
|
|
|
|
static void DeadlockTest(SpinLock* spinlock, int num_spinners) {
|
|
Notification locked;
|
|
BlockingCounter counter(num_spinners);
|
|
std::vector<std::thread> threads;
|
|
|
|
threads.push_back(
|
|
std::thread(Helper::LockThenWait, &locked, spinlock, &counter));
|
|
for (int i = 0; i < num_spinners; ++i) {
|
|
threads.push_back(
|
|
std::thread(Helper::NotifyThenLock, &locked, spinlock, &counter));
|
|
}
|
|
|
|
for (auto& thread : threads) {
|
|
thread.join();
|
|
}
|
|
}
|
|
};
|
|
|
|
SpinLock stack_cooperative_spinlock(
|
|
base_internal::SCHEDULE_COOPERATIVE_AND_KERNEL);
|
|
SpinLock stack_noncooperative_spinlock(base_internal::SCHEDULE_KERNEL_ONLY);
|
|
Helper::DeadlockTest(&stack_cooperative_spinlock,
|
|
base_internal::NumCPUs() * 2);
|
|
Helper::DeadlockTest(&stack_noncooperative_spinlock,
|
|
base_internal::NumCPUs() * 2);
|
|
Helper::DeadlockTest(&static_cooperative_spinlock,
|
|
base_internal::NumCPUs() * 2);
|
|
Helper::DeadlockTest(&static_noncooperative_spinlock,
|
|
base_internal::NumCPUs() * 2);
|
|
}
|
|
|
|
} // namespace
|
|
} // namespace base_internal
|
|
ABSL_NAMESPACE_END
|
|
} // namespace absl
|