The sparse-map library is a C++ implementation of a memory efficient hash map and hash set based on tsl::sparse_map. We added support for fancy pointers. It uses open-addressing with sparse quadratic probing. The goal of the library is to be the most memory efficient possible, even at low load factor, while keeping reasonable performances. You can find an article of Stephen Merity which explains the idea behind google::sparse_hash_map
and this project.
Four classes are provided: dice::sparse_map::sparse_map
, dice::sparse_map::sparse_set
, dice::sparse_map::sparse_pg_map
and dice::sparse_map::sparse_pg_set
. The first two are faster and use a power of two growth policy, the last two use a prime growth policy instead and are able to cope better with a poor hash function. Use the prime version if there is a chance of repeating patterns in the lower bits of your hash (e.g. you are storing pointers with an identity hash function). See GrowthPolicy for details.
A benchmark of dice::sparse_map::sparse_map
against other hash maps may be found here. The benchmark, in its additional tests page, notably includes google::sparse_hash_map
and spp::sparse_hash_map
to which dice::sparse_map::sparse_map
is an alternative. This page also gives some advices on which hash table structure you should try for your use case (useful if you are a bit lost with the multiple hash tables implementations in the tsl
namespace).
- Header-only library, just add the include directory to your include path and you are ready to go. If you use CMake, you can also use the
dice::sparse_map::sparse_map
exported target from the CMakeLists.txt. - Memory efficient while keeping good lookup speed, see the benchmark for some numbers.
- Support for heterogeneous lookups allowing the usage of
find
with a type different thanKey
(e.g. if you have a map that usesstd::unique_ptr<foo>
as key, you can use afoo*
or astd::uintptr_t
as key parameter tofind
without constructing astd::unique_ptr<foo>
, see example). - No need to reserve any sentinel value from the keys.
- If the hash is known before a lookup, it is possible to pass it as parameter to speed-up the lookup (see
precalculated_hash
parameter in API). - Support for efficient serialization and deserialization (see example and the
serialize/deserialize
methods in the API for details). - Possibility to control the balance between insertion speed and memory usage with the
Sparsity
template parameter. A high sparsity means less memory but longer insertion times, and vice-versa for low sparsity. The default medium sparsity offers a good compromise (see API for details). For reference, with simple 64 bits integers as keys and values, a low sparsity offers ~15% faster insertions times but uses ~12% more memory. Nothing change regarding lookup speed. - API closely similar to
std::unordered_map
andstd::unordered_set
.
dice::sparse_map::sparse_map
tries to have an interface similar to std::unordered_map
, but some differences exist.
- By default only the basic exception safety is guaranteed which mean that, in case of exception, all resources used by the hash map will be freed (no memory leaks) but the hash map may end-up in an undefined state (undefined here means that some elements may be missing). This can ONLY happen on rehash (either on insert or if
rehash
is called explicitly) and will occur if theAllocator
can't allocate memory (std::bad_alloc
) or if the copy constructor (when a nothrow move constructor is not available) throws and exception. This can be avoided by callingreserve
beforehand. It is the same guarantee that the one provided bygoogle::sparse_hash_map
andspp::sparse_hash_map
which don't provide the strong exception guarantee. For more information and if you need the strong exception guarantee, check theExceptionSafety
template parameter (see API for details). - Iterator invalidation doesn't behave in the same way, any operation modifying the hash table invalidate them (see API for details).
- References and pointers to keys or values in the map are invalidated in the same way as iterators to these keys-values.
- For iterators,
operator*()
andoperator->()
return a reference and a pointer toconst std::pair<Key, T>
instead ofstd::pair<const Key, T>
making the valueT
not modifiable. To modify the value you have to call thevalue()
method of the iterator to get a mutable reference. Example:
dice::sparse_map::sparse_map<int, int> map = {{1, 1}, {2, 1}, {3, 1}};
for(auto it = map.begin(); it != map.end(); ++it) {
//it->second = 2; // Illegal
it.value() = 2; // Ok
}
- Move-only types must have a nothrow move constructor.
- No support for some buckets related methods (like
bucket_size
,bucket
, ...).
These differences also apply between std::unordered_set
and dice::sparse_map::sparse_set
.
Thread-safety guarantees are the same as std::unordered_map/set
(i.e. possible to have multiple readers with no writer).
The library relies heavily on the popcount operation.
With Clang and GCC, the library uses the __builtin_popcount
function which will use the fast CPU instruction POPCNT
when the library is compiled with -mpopcnt
. Using the POPCNT
instruction offers an improvement of ~15% to ~30% on lookups. So if you are compiling your code for a specific architecture that support the operation, don't forget the -mpopcnt
(or -march=native
) flag of your compiler.
On Windows with MSVC, the detection is done at runtime.
Make sure that your key Key
and potential value T
have a noexcept
move constructor. The library will work without it but insertions will be much slower if the copy constructor is expensive (the structure often needs to move some values around on insertion).
The library supports multiple growth policies through the GrowthPolicy
template parameter. Three policies are provided by the library but you can easily implement your own if needed.
- dice::sparse_map::sh::power_of_two_growth_policy. Default policy used by
dice::sparse_map::sparse_map/set
. This policy keeps the size of the bucket array of the hash table to a power of two. This constraint allows the policy to avoid the usage of the slow modulo operation to map a hash to a bucket, instead ofhash % 2n
, it useshash & (2n - 1)
(see fast modulo). Fast but this may cause a lot of collisions with a poor hash function as the modulo with a power of two only masks the most significant bits in the end. - dice::sparse_map::sh::prime_growth_policy. Default policy used by
dice::sparse_map::sparse_pg_map/set
. The policy keeps the size of the bucket array of the hash table to a prime number. When mapping a hash to a bucket, using a prime number as modulo will result in a better distribution of the hash across the buckets even with a poor hash function. To allow the compiler to optimize the modulo operation, the policy use a lookup table with constant primes modulos (see API for details). Slower thandice::sparse_map::sh::power_of_two_growth_policy
but more secure. - dice::sparse_map::sh::mod_growth_policy. The policy grows the map by a customizable growth factor passed in parameter. It then just use the modulo operator to map a hash to a bucket. Slower but more flexible.
To implement your own policy, you have to implement the following interface.
struct custom_policy {
// Called on hash table construction and rehash, min_bucket_count_in_out is the minimum buckets
// that the hash table needs. The policy can change it to a higher number of buckets if needed
// and the hash table will use this value as bucket count. If 0 bucket is asked, then the value
// must stay at 0.
explicit custom_policy(std::size_t& min_bucket_count_in_out);
// Return the bucket [0, bucket_count()) to which the hash belongs.
// If bucket_count() is 0, it must always return 0.
std::size_t bucket_for_hash(std::size_t hash) const noexcept;
// Return the number of buckets that should be used on next growth
std::size_t next_bucket_count() const;
// Return the maximum number of buckets supported by the policy.
std::size_t max_bucket_count() const;
// Reset the growth policy as if it was created with a bucket count of 0.
// After a clear, the policy must always return 0 when bucket_for_hash is called.
void clear() noexcept;
}
To use sparse-map, just add the include directory to your include path. It is a header-only library.
If you use CMake, you can also use the dice::sparse_map::sparse_map
exported target from the CMakeLists.txt with target_link_libraries
.
# Example where the sparse-map project is stored in a third-party directory
add_subdirectory(third-party/sparse-map)
target_link_libraries(your_target PRIVATE dice::sparse_map)
If the project has been installed through make install
, you can also use find_package(tsl-sparse-map REQUIRED)
instead of add_subdirectory
.
The code should work with any C++11 standard-compliant compiler and has been tested with GCC 4.8.4, Clang 3.5.0 and Visual Studio 2015.
To run the tests you will need the Boost Test library and CMake.
git clone https://github.com/Tessil/sparse-map.git
cd sparse-map/tests
mkdir build
cd build
cmake ..
cmake --build .
./tsl_sparse_map_tests
The API can be found here.
All methods are not documented yet, but they replicate the behaviour of the ones in std::unordered_map
and std::unordered_set
, except if specified otherwise.
#include <cstdint>
#include <iostream>
#include <string>
#include <dice/sparse-map/sparse_map.hpp>
#include <dice/sparse-map/sparse_set.hpp>
int main() {
dice::sparse_map::sparse_map<std::string, int> map = {{"a", 1}, {"b", 2}};
map["c"] = 3;
map["d"] = 4;
map.insert({"e", 5});
map.erase("b");
for(auto it = map.begin(); it != map.end(); ++it) {
//it->second += 2; // Not valid.
it.value() += 2;
}
// {d, 6} {a, 3} {e, 7} {c, 5}
for(const auto& key_value : map) {
std::cout << "{" << key_value.first << ", " << key_value.second << "}" << std::endl;
}
if(map.find("a") != map.end()) {
std::cout << "Found \"a\"." << std::endl;
}
const std::size_t precalculated_hash = std::hash<std::string>()("a");
// If we already know the hash beforehand, we can pass it as argument to speed-up the lookup.
if(map.find("a", precalculated_hash) != map.end()) {
std::cout << "Found \"a\" with hash " << precalculated_hash << "." << std::endl;
}
dice::sparse_map::sparse_set<int> set;
set.insert({1, 9, 0});
set.insert({2, -1, 9});
// {0} {1} {2} {9} {-1}
for(const auto& key : set) {
std::cout << "{" << key << "}" << std::endl;
}
}
Heterogeneous overloads allow the usage of other types than Key
for lookup and erase operations as long as the used types are hashable and comparable to Key
.
To activate the heterogeneous overloads in dice::sparse_map::sparse_map/set
, the qualified-id KeyEqual::is_transparent
must be valid. It works the same way as for std::map::find
. You can either use std::equal_to<>
or define your own function object.
Both KeyEqual
and Hash
will need to be able to deal with the different types.
#include <functional>
#include <iostream>
#include <string>
#include <dice/sparse-map/sparse_map.hpp>
struct employee {
employee(int id, std::string name) : m_id(id), m_name(std::move(name)) {
}
// Either we include the comparators in the class and we use `std::equal_to<>`...
friend bool operator==(const employee& empl, int empl_id) {
return empl.m_id == empl_id;
}
friend bool operator==(int empl_id, const employee& empl) {
return empl_id == empl.m_id;
}
friend bool operator==(const employee& empl1, const employee& empl2) {
return empl1.m_id == empl2.m_id;
}
int m_id;
std::string m_name;
};
// ... or we implement a separate class to compare employees.
struct equal_employee {
using is_transparent = void;
bool operator()(const employee& empl, int empl_id) const {
return empl.m_id == empl_id;
}
bool operator()(int empl_id, const employee& empl) const {
return empl_id == empl.m_id;
}
bool operator()(const employee& empl1, const employee& empl2) const {
return empl1.m_id == empl2.m_id;
}
};
struct hash_employee {
std::size_t operator()(const employee& empl) const {
return std::hash<int>()(empl.m_id);
}
std::size_t operator()(int id) const {
return std::hash<int>()(id);
}
};
int main() {
// Use std::equal_to<> which will automatically deduce and forward the parameters
dice::sparse_map::sparse_map<employee, int, hash_employee, std::equal_to<>> map;
map.insert({employee(1, "John Doe"), 2001});
map.insert({employee(2, "Jane Doe"), 2002});
map.insert({employee(3, "John Smith"), 2003});
// John Smith 2003
auto it = map.find(3);
if(it != map.end()) {
std::cout << it->first.m_name << " " << it->second << std::endl;
}
map.erase(1);
// Use a custom KeyEqual which has an is_transparent member type
dice::sparse_map::sparse_map<employee, int, hash_employee, equal_employee> map2;
map2.insert({employee(4, "Johnny Doe"), 2004});
// 2004
std::cout << map2.at(4) << std::endl;
}
The library provides an efficient way to serialize and deserialize a map or a set so that it can be saved to a file or send through the network. To do so, it requires the user to provide a function object for both serialization and deserialization.
struct serializer {
// Must support the following types for U: std::uint64_t, float
// and std::pair<Key, T> if a map is used or Key for a set.
template<typename U>
void operator()(const U& value);
};
struct deserializer {
// Must support the following types for U: std::uint64_t, float
// and std::pair<Key, T> if a map is used or Key for a set.
template<typename U>
U operator()();
};
Note that the implementation leaves binary compatibility (endianness, float binary representation, size of int, ...) of the types it serializes/deserializes in the hands of the provided function objects if compatibility is required.
More details regarding the serialize
and deserialize
methods can be found in the API.
#include <cassert>
#include <cstdint>
#include <fstream>
#include <type_traits>
#include <dice/sparse-map/sparse_map.hpp>
class serializer {
public:
serializer(const char* file_name) {
m_ostream.exceptions(m_ostream.badbit | m_ostream.failbit);
m_ostream.open(file_name, std::ios::binary);
}
template<class T,
typename std::enable_if<std::is_arithmetic<T>::value>::type* = nullptr>
void operator()(const T& value) {
m_ostream.write(reinterpret_cast<const char*>(&value), sizeof(T));
}
void operator()(const std::pair<std::int64_t, std::int64_t>& value) {
(*this)(value.first);
(*this)(value.second);
}
private:
std::ofstream m_ostream;
};
class deserializer {
public:
deserializer(const char* file_name) {
m_istream.exceptions(m_istream.badbit | m_istream.failbit | m_istream.eofbit);
m_istream.open(file_name, std::ios::binary);
}
template<class T>
T operator()() {
T value;
deserialize(value);
return value;
}
private:
template<class T,
typename std::enable_if<std::is_arithmetic<T>::value>::type* = nullptr>
void deserialize(T& value) {
m_istream.read(reinterpret_cast<char*>(&value), sizeof(T));
}
void deserialize(std::pair<std::int64_t, std::int64_t>& value) {
deserialize(value.first);
deserialize(value.second);
}
private:
std::ifstream m_istream;
};
int main() {
const dice::sparse_map::sparse_map<std::int64_t, std::int64_t> map = {{1, -1}, {2, -2}, {3, -3}, {4, -4}};
const char* file_name = "sparse_map.data";
{
serializer serial(file_name);
map.serialize(serial);
}
{
deserializer dserial(file_name);
auto map_deserialized = dice::sparse_map::sparse_map<std::int64_t, std::int64_t>::deserialize(dserial);
assert(map == map_deserialized);
}
{
deserializer dserial(file_name);
/**
* If the serialized and deserialized map are hash compatibles (see conditions in API),
* setting the argument to true speed-up the deserialization process as we don't have
* to recalculate the hash of each key. We also know how much space each bucket needs.
*/
const bool hash_compatible = true;
auto map_deserialized =
dice::sparse_map::sparse_map<std::int64_t, std::int64_t>::deserialize(dserial, hash_compatible);
assert(map == map_deserialized);
}
}
It's possible to use a serialization library to avoid the boilerplate.
The following example uses Boost Serialization with the Boost zlib compression stream to reduce the size of the resulting serialized file. The example requires C++20 due to the usage of the template parameter list syntax in lambdas, but it can be adapted to less recent versions.
#include <boost/archive/binary_iarchive.hpp>
#include <boost/archive/binary_oarchive.hpp>
#include <boost/iostreams/filter/zlib.hpp>
#include <boost/iostreams/filtering_stream.hpp>
#include <boost/serialization/split_free.hpp>
#include <boost/serialization/utility.hpp>
#include <cassert>
#include <cstdint>
#include <fstream>
#include <dice/sparse-map/sparse_map.hpp>
namespace boost { namespace serialization {
template<class Archive, class Key, class T>
void serialize(Archive & ar, dice::sparse_map::sparse_map<Key, T>& map, const unsigned int version) {
split_free(ar, map, version);
}
template<class Archive, class Key, class T>
void save(Archive & ar, const dice::sparse_map::sparse_map<Key, T>& map, const unsigned int /*version*/) {
auto serializer = [&ar](const auto& v) { ar & v; };
map.serialize(serializer);
}
template<class Archive, class Key, class T>
void load(Archive & ar, dice::sparse_map::sparse_map<Key, T>& map, const unsigned int /*version*/) {
auto deserializer = [&ar]<typename U>() { U u; ar & u; return u; };
map = dice::sparse_map::sparse_map<Key, T>::deserialize(deserializer);
}
}}
int main() {
dice::sparse_map::sparse_map<std::int64_t, std::int64_t> map = {{1, -1}, {2, -2}, {3, -3}, {4, -4}};
const char* file_name = "sparse_map.data";
{
std::ofstream ofs;
ofs.exceptions(ofs.badbit | ofs.failbit);
ofs.open(file_name, std::ios::binary);
boost::iostreams::filtering_ostream fo;
fo.push(boost::iostreams::zlib_compressor());
fo.push(ofs);
boost::archive::binary_oarchive oa(fo);
oa << map;
}
{
std::ifstream ifs;
ifs.exceptions(ifs.badbit | ifs.failbit | ifs.eofbit);
ifs.open(file_name, std::ios::binary);
boost::iostreams::filtering_istream fi;
fi.push(boost::iostreams::zlib_decompressor());
fi.push(ifs);
boost::archive::binary_iarchive ia(fi);
dice::sparse_map::sparse_map<std::int64_t, std::int64_t> map_deserialized;
ia >> map_deserialized;
assert(map == map_deserialized);
}
}
The code is licensed under the MIT license, see the LICENSE files (1, 2) for details.