1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52
|
#define THIS_HASH_MAP parallel_flat_hash_map
#define THIS_TEST_NAME ParallelFlatHashMap
#include <thread>
#include "parallel_hash_map_test.cc"
#if PHMAP_HAVE_SHARED_MUTEX
#include <shared_mutex>
template <typename K> using HashEqual = phmap::priv::hash_default_eq<K>;
template <typename V> using HashFn = phmap::priv::hash_default_hash<V>;
template <typename K> using Allocator = phmap::priv::Allocator<K>;
template <typename K, typename V, size_t N>
using parallel_flat_hash_map =
phmap::parallel_flat_hash_map<K, V, HashFn<K>, HashEqual<K>,
Allocator<phmap::priv::Pair<K, V>>, N,
std::shared_mutex>;
using Table = parallel_flat_hash_map<int, int, 10>;
TEST(THIS_TEST_NAME, ConcurrencyCheck) {
static constexpr int THREADS = 10;
static constexpr int EPOCH = 1000;
static constexpr int KEY = 12345;
auto Incr = [](Table *table) {
auto exist_fn = [](typename Table::value_type &value) { value.second += 1; };
auto emplace_fn = [](const typename Table::constructor &ctor) {
ctor(KEY, 1);
};
for (int i = 0; i < EPOCH; ++i) {
(void)table->lazy_emplace_l(KEY, exist_fn, emplace_fn);
}
};
Table table;
std::vector<std::thread> threads;
threads.reserve(THREADS);
for (int i = 0; i < THREADS; ++i) {
threads.emplace_back([&]() { Incr(&table); });
}
for (auto &thread : threads) {
thread.join();
}
EXPECT_EQ(table[KEY], 10000);
}
#endif
|