diff --git a/Cargo.toml b/Cargo.toml index d733140c..d0fa075c 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -49,6 +49,8 @@ sanakirja-core = "=1.4.1" rocksdb = { version = "0.22.0", default-features = false, features = ["lz4"] } fjall = "2.6" comfy-table = "7.0.1" +canopydb = "0.2" +env_logger = "0.11" [target.'cfg(target_os = "linux")'.dev-dependencies] io-uring = "0.6.2" diff --git a/benches/common.rs b/benches/common.rs index 0ccab252..38047378 100644 --- a/benches/common.rs +++ b/benches/common.rs @@ -1083,3 +1083,127 @@ impl BenchInserter for FjallBenchInserter<'_, '_> { Ok(()) } } + +pub struct CanopydbBenchDatabase<'a> { + db: &'a canopydb::Database, +} + +impl<'a> CanopydbBenchDatabase<'a> { + #[allow(dead_code)] + pub fn new(db: &'a canopydb::Database) -> Self { + Self { db } + } +} + +impl<'a> BenchDatabase for CanopydbBenchDatabase<'a> { + type W<'db>= CanopydbBenchWriteTransaction<'a> where Self: 'db; + type R<'db> = CanopydbBenchReadTransaction where Self: 'db; + + fn db_type_name() -> &'static str { + "canopydb" + } + + fn write_transaction(&self) -> Self::W<'_> { + CanopydbBenchWriteTransaction { + db: self.db, + txn: self.db.begin_write().unwrap(), + } + } + + fn read_transaction(&self) -> Self::R<'_> { + CanopydbBenchReadTransaction { + txn: self.db.begin_read().unwrap(), + } + } + + fn compact(&mut self) -> bool { + self.db.compact().unwrap(); + true + } +} + +pub struct CanopydbBenchWriteTransaction<'db> { + db: &'db canopydb::Database, + txn: canopydb::WriteTransaction, +} + +impl BenchWriteTransaction for CanopydbBenchWriteTransaction<'_> { + type W<'txn> = CanopydbBenchInserter<'txn> where Self: 'txn; + + fn get_inserter(&mut self) -> Self::W<'_> { + CanopydbBenchInserter { + tree: self.txn.get_or_create_tree(b"default").unwrap(), + } + } + + fn commit(self) -> Result<(), ()> { + self.txn.commit().map_err(|_| ())?; + self.db.sync().map_err(|_| ()) + } +} + +pub struct CanopydbBenchInserter<'txn> { + tree: canopydb::Tree<'txn>, +} + +impl BenchInserter for CanopydbBenchInserter<'_> { + fn insert(&mut self, key: &[u8], value: &[u8]) -> Result<(), ()> { + self.tree.insert(key, value).map(|_| ()).map_err(|_| ()) + } + + fn remove(&mut self, key: &[u8]) -> Result<(), ()> { + self.tree.delete(key).map(|_| ()).map_err(|_| ()) + } +} + +pub struct CanopydbBenchReadTransaction { + txn: canopydb::ReadTransaction, +} + +impl BenchReadTransaction for CanopydbBenchReadTransaction { + type T<'txn> = CanopydbBenchReader<'txn> where Self: 'txn; + + fn get_reader(&self) -> Self::T<'_> { + CanopydbBenchReader { + tree: self.txn.get_tree(b"default").unwrap().unwrap(), + } + } +} + +pub struct CanopydbBenchReader<'txn> { + tree: canopydb::Tree<'txn>, +} + +impl BenchReader for CanopydbBenchReader<'_> { + type Output<'out> = canopydb::Bytes where Self: 'out; + type Iterator<'out> = CanopydbBenchIterator<'out> where Self: 'out; + + fn get(&self, key: &[u8]) -> Option> { + self.tree.get(key).unwrap() + } + + fn range_from<'a>(&'a self, key: &'a [u8]) -> Self::Iterator<'a> { + CanopydbBenchIterator { + iter: self.tree.range(key..).unwrap(), + } + } + + fn len(&self) -> u64 { + self.tree.len() + } +} + +pub struct CanopydbBenchIterator<'tree> { + iter: canopydb::RangeIter<'tree>, +} + +impl BenchIterator for CanopydbBenchIterator<'_> { + type Output<'out> = canopydb::Bytes where Self: 'out; + + fn next(&mut self) -> Option<(Self::Output<'_>, Self::Output<'_>)> { + self.iter.next().map(|x| { + let x = x.unwrap(); + (x.0, x.1) + }) + } +} diff --git a/benches/int_benchmark.rs b/benches/int_benchmark.rs index 63ef746f..095ec337 100644 --- a/benches/int_benchmark.rs +++ b/benches/int_benchmark.rs @@ -79,7 +79,19 @@ fn main() { let rocksdb_results = { let tmpfile: TempDir = tempfile::tempdir_in(current_dir().unwrap()).unwrap(); - let db = rocksdb::OptimisticTransactionDB::open_default(tmpfile.path()).unwrap(); + + let mut bb = rocksdb::BlockBasedOptions::default(); + bb.set_block_cache(&rocksdb::Cache::new_lru_cache(4 * 1_024 * 1_024 * 1_024)); + bb.set_bloom_filter(10.0, false); + + let mut opts = rocksdb::Options::default(); + opts.set_block_based_table_factory(&bb); + opts.create_if_missing(true); + opts.increase_parallelism( + std::thread::available_parallelism().map_or(1, |n| n.get()) as i32 + ); + + let db = rocksdb::OptimisticTransactionDB::open(&opts, tmpfile.path()).unwrap(); let table = RocksdbBenchDatabase::new(&db); benchmark(table) }; @@ -99,6 +111,15 @@ fn main() { benchmark(table) }; + let canopydb_results = { + let tmpfile: TempDir = tempfile::tempdir_in(current_dir().unwrap()).unwrap(); + let mut env_opts = canopydb::EnvOptions::new(tmpfile.path()); + env_opts.page_cache_size = 4 * 1024 * 1024 * 1024; + let db = canopydb::Database::with_options(env_opts, Default::default()).unwrap(); + let db_bench = CanopydbBenchDatabase::new(&db); + benchmark(db_bench) + }; + let mut rows = Vec::new(); for (benchmark, _duration) in &redb_results { @@ -111,6 +132,7 @@ fn main() { rocksdb_results, sled_results, sanakirja_results, + canopydb_results, ] { for (i, (_benchmark, duration)) in results.iter().enumerate() { rows[i].push(format!("{}ms", duration.as_millis())); @@ -119,7 +141,15 @@ fn main() { let mut table = comfy_table::Table::new(); table.set_width(100); - table.set_header(["", "redb", "lmdb", "rocksdb", "sled", "sanakirja"]); + table.set_header([ + "", + "redb", + "lmdb", + "rocksdb", + "sled", + "sanakirja", + "canopydb", + ]); for row in rows { table.add_row(row); } diff --git a/benches/large_values_benchmark.rs b/benches/large_values_benchmark.rs index 6fc6ba98..424b4fb2 100644 --- a/benches/large_values_benchmark.rs +++ b/benches/large_values_benchmark.rs @@ -6,7 +6,7 @@ use tempfile::{NamedTempFile, TempDir}; mod common; use common::*; -use rand::Rng; +use rand::RngCore; use std::time::{Duration, Instant}; const ELEMENTS: usize = 1_000_000; @@ -16,8 +16,10 @@ fn random_data(count: usize, key_size: usize, value_size: usize) -> Vec<(Vec let mut pairs = vec![]; for _ in 0..count { - let key: Vec = (0..key_size).map(|_| rand::rng().random()).collect(); - let value: Vec = (0..value_size).map(|_| rand::rng().random()).collect(); + let mut key = vec![0; key_size]; + rand::rng().fill_bytes(&mut key); + let mut value = vec![0; value_size]; + rand::rng().fill_bytes(&mut value); pairs.push((key, value)); } @@ -69,6 +71,8 @@ fn benchmark(db: T) -> Vec<(&'static str, Duration)> { } fn main() { + let _ = env_logger::try_init(); + let redb_latency_results = { let tmpfile: NamedTempFile = NamedTempFile::new_in(current_dir().unwrap()).unwrap(); let mut db = redb::Database::builder().create(tmpfile.path()).unwrap(); @@ -90,7 +94,19 @@ fn main() { let rocksdb_results = { let tmpfile: TempDir = tempfile::tempdir_in(current_dir().unwrap()).unwrap(); - let db = rocksdb::OptimisticTransactionDB::open_default(tmpfile.path()).unwrap(); + + let mut bb = rocksdb::BlockBasedOptions::default(); + bb.set_block_cache(&rocksdb::Cache::new_lru_cache(4 * 1_024 * 1_024 * 1_024)); + bb.set_bloom_filter(10.0, false); + + let mut opts = rocksdb::Options::default(); + opts.set_block_based_table_factory(&bb); + opts.create_if_missing(true); + opts.increase_parallelism( + std::thread::available_parallelism().map_or(1, |n| n.get()) as i32 + ); + + let db = rocksdb::OptimisticTransactionDB::open(&opts, tmpfile.path()).unwrap(); let table = RocksdbBenchDatabase::new(&db); benchmark(table) }; @@ -102,6 +118,15 @@ fn main() { benchmark(table) }; + let canopydb_results = { + let tmpfile: TempDir = tempfile::tempdir_in(current_dir().unwrap()).unwrap(); + let mut env_opts = canopydb::EnvOptions::new(tmpfile.path()); + env_opts.page_cache_size = 4 * 1024 * 1024 * 1024; + let db = canopydb::Database::with_options(env_opts, Default::default()).unwrap(); + let db_bench = CanopydbBenchDatabase::new(&db); + benchmark(db_bench) + }; + let mut rows = Vec::new(); for (benchmark, _duration) in &redb_latency_results { @@ -113,6 +138,7 @@ fn main() { lmdb_results, rocksdb_results, sled_results, + canopydb_results, ] { for (i, (_benchmark, duration)) in results.iter().enumerate() { rows[i].push(format!("{}ms", duration.as_millis())); @@ -121,7 +147,7 @@ fn main() { let mut table = comfy_table::Table::new(); table.set_width(100); - table.set_header(["", "redb", "lmdb", "rocksdb", "sled"]); + table.set_header(["", "redb", "lmdb", "rocksdb", "sled", "canopydb"]); for row in rows { table.add_row(row); } diff --git a/benches/lmdb_benchmark.rs b/benches/lmdb_benchmark.rs index dd114aeb..188dc6c6 100644 --- a/benches/lmdb_benchmark.rs +++ b/benches/lmdb_benchmark.rs @@ -1,5 +1,4 @@ use std::env::current_dir; -use std::mem::size_of; use std::path::Path; use std::sync::Arc; use std::{fs, process, thread}; @@ -10,47 +9,27 @@ use common::*; use std::time::{Duration, Instant}; -const ITERATIONS: usize = 2; -const ELEMENTS: usize = 1_000_000; +const READ_ITERATIONS: usize = 2; +const BULK_ELEMENTS: usize = 1_000_000; +const INDIVIDUAL_WRITES: usize = 1_000; +const BATCH_WRITES: usize = 100; +const BATCH_SIZE: usize = 1000; +const SCAN_ITERATIONS: usize = 2; +const NUM_READS: usize = 1_000_000; +const NUM_SCANS: usize = 500_000; +const SCAN_LEN: usize = 10; const KEY_SIZE: usize = 24; const VALUE_SIZE: usize = 150; const RNG_SEED: u64 = 3; -const CACHE_SIZE: usize = 4 * 1_024 * 1_024 * 1_024; - -fn fill_slice(slice: &mut [u8], rng: &mut fastrand::Rng) { - let mut i = 0; - while i + size_of::() < slice.len() { - let tmp = rng.u128(..); - slice[i..(i + size_of::())].copy_from_slice(&tmp.to_le_bytes()); - i += size_of::() - } - if i + size_of::() < slice.len() { - let tmp = rng.u64(..); - slice[i..(i + size_of::())].copy_from_slice(&tmp.to_le_bytes()); - i += size_of::() - } - if i + size_of::() < slice.len() { - let tmp = rng.u32(..); - slice[i..(i + size_of::())].copy_from_slice(&tmp.to_le_bytes()); - i += size_of::() - } - if i + size_of::() < slice.len() { - let tmp = rng.u16(..); - slice[i..(i + size_of::())].copy_from_slice(&tmp.to_le_bytes()); - i += size_of::() - } - if i + size_of::() < slice.len() { - slice[i] = rng.u8(..); - } -} +const CACHE_SIZE: usize = 4 * 1_024 * 1_024 * 1_024; // 4GB /// Returns pairs of key, value fn random_pair(rng: &mut fastrand::Rng) -> ([u8; KEY_SIZE], Vec) { let mut key = [0u8; KEY_SIZE]; - fill_slice(&mut key, rng); + rng.fill(&mut key); let mut value = vec![0u8; VALUE_SIZE]; - fill_slice(&mut value, rng); + rng.fill(&mut value); (key, value) } @@ -82,7 +61,7 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, let mut txn = db.write_transaction(); let mut inserter = txn.get_inserter(); { - for _ in 0..ELEMENTS { + for _ in 0..BULK_ELEMENTS { let (key, value) = random_pair(&mut rng); inserter.insert(&key, &value).unwrap(); } @@ -95,15 +74,14 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, println!( "{}: Bulk loaded {} items in {}ms", T::db_type_name(), - ELEMENTS, + BULK_ELEMENTS, duration.as_millis() ); results.push(("bulk load".to_string(), ResultType::Duration(duration))); let start = Instant::now(); - let writes = 100; { - for _ in 0..writes { + for _ in 0..INDIVIDUAL_WRITES { let mut txn = db.write_transaction(); let mut inserter = txn.get_inserter(); let (key, value) = random_pair(&mut rng); @@ -118,7 +96,7 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, println!( "{}: Wrote {} individual items in {}ms", T::db_type_name(), - writes, + INDIVIDUAL_WRITES, duration.as_millis() ); results.push(( @@ -127,12 +105,11 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, )); let start = Instant::now(); - let batch_size = 1000; { - for _ in 0..writes { + for _ in 0..BATCH_WRITES { let mut txn = db.write_transaction(); let mut inserter = txn.get_inserter(); - for _ in 0..batch_size { + for _ in 0..BATCH_SIZE { let (key, value) = random_pair(&mut rng); inserter.insert(&key, &value).unwrap(); } @@ -144,33 +121,34 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, let end = Instant::now(); let duration = end - start; println!( - "{}: Wrote {} x {} items in {}ms", + "{}: Wrote {} batches of {} items in {}ms", T::db_type_name(), - writes, - batch_size, + BATCH_WRITES, + BATCH_SIZE, duration.as_millis() ); results.push(("batch writes".to_string(), ResultType::Duration(duration))); + let elements = BULK_ELEMENTS + INDIVIDUAL_WRITES + BATCH_SIZE * BATCH_WRITES; let txn = db.read_transaction(); { { let start = Instant::now(); let len = txn.get_reader().len(); - assert_eq!(len, ELEMENTS as u64 + 100_000 + 100); + assert_eq!(len, elements as u64); let end = Instant::now(); let duration = end - start; println!("{}: len() in {}ms", T::db_type_name(), duration.as_millis()); results.push(("len()".to_string(), ResultType::Duration(duration))); } - for _ in 0..ITERATIONS { + for _ in 0..READ_ITERATIONS { let mut rng = make_rng(); let start = Instant::now(); let mut checksum = 0u64; let mut expected_checksum = 0u64; let reader = txn.get_reader(); - for _ in 0..ELEMENTS { + for _ in 0..NUM_READS { let (key, value) = random_pair(&mut rng); let result = reader.get(&key).unwrap(); checksum += result.as_ref()[0] as u64; @@ -182,22 +160,21 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, println!( "{}: Random read {} items in {}ms", T::db_type_name(), - ELEMENTS, + NUM_READS, duration.as_millis() ); results.push(("random reads".to_string(), ResultType::Duration(duration))); } - for _ in 0..ITERATIONS { + for _ in 0..SCAN_ITERATIONS { let mut rng = make_rng(); let start = Instant::now(); let reader = txn.get_reader(); let mut value_sum = 0; - let num_scan = 10; - for _ in 0..ELEMENTS { + for _ in 0..NUM_SCANS { let (key, _value) = random_pair(&mut rng); let mut iter = reader.range_from(&key); - for _ in 0..num_scan { + for _ in 0..SCAN_LEN { if let Some((_, value)) = iter.next() { value_sum += value.as_ref()[0]; } else { @@ -209,9 +186,10 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, let end = Instant::now(); let duration = end - start; println!( - "{}: Random range read {} elements in {}ms", + "{}: Random range read {} x {} elements in {}ms", T::db_type_name(), - ELEMENTS * num_scan, + NUM_SCANS, + SCAN_LEN, duration.as_millis() ); results.push(( @@ -223,19 +201,23 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, drop(txn); for num_threads in [4, 8, 16, 32] { - let mut rngs = make_rng_shards(num_threads, ELEMENTS); + let barrier = Arc::new(std::sync::Barrier::new(num_threads)); + let mut rngs = make_rng_shards(num_threads, elements); let start = Instant::now(); thread::scope(|s| { for _ in 0..num_threads { + let barrier = barrier.clone(); let db2 = db.clone(); - let mut rng = rngs.pop().unwrap(); + let rng = rngs.pop().unwrap(); s.spawn(move || { + barrier.wait(); let txn = db2.read_transaction(); let mut checksum = 0u64; let mut expected_checksum = 0u64; let reader = txn.get_reader(); - for _ in 0..(ELEMENTS / num_threads) { + let mut rng = rng.clone(); + for _ in 0..(elements / num_threads) { let (key, value) = random_pair(&mut rng); let result = reader.get(&key).unwrap(); checksum += result.as_ref()[0] as u64; @@ -252,7 +234,7 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, "{}: Random read ({} threads) {} items in {}ms", T::db_type_name(), num_threads, - ELEMENTS, + elements, duration.as_millis() ); results.push(( @@ -262,7 +244,7 @@ fn benchmark(db: T, path: &Path) -> Vec<(String, } let start = Instant::now(); - let deletes = ELEMENTS / 2; + let deletes = elements / 2; { let mut rng = make_rng(); let mut txn = db.write_transaction(); @@ -351,6 +333,7 @@ impl std::fmt::Display for ResultType { } fn main() { + let _ = env_logger::try_init(); let tmpdir = current_dir().unwrap().join(".benchmark"); fs::create_dir(&tmpdir).unwrap(); @@ -388,10 +371,14 @@ fn main() { let mut bb = rocksdb::BlockBasedOptions::default(); bb.set_block_cache(&rocksdb::Cache::new_lru_cache(CACHE_SIZE)); + bb.set_bloom_filter(10.0, false); let mut opts = rocksdb::Options::default(); opts.set_block_based_table_factory(&bb); opts.create_if_missing(true); + opts.increase_parallelism( + std::thread::available_parallelism().map_or(1, |n| n.get()) as i32 + ); let db = rocksdb::OptimisticTransactionDB::open(&opts, tmpfile.path()).unwrap(); let table = RocksdbBenchDatabase::new(&db); @@ -433,6 +420,15 @@ fn main() { benchmark(table, tmpfile.path()) }; + let canopydb_results = { + let tmpdir = tempfile::tempdir_in(&tmpdir).unwrap(); + let mut env_opts = canopydb::EnvOptions::new(tmpdir.path()); + env_opts.page_cache_size = CACHE_SIZE; + let db = canopydb::Database::with_options(env_opts, Default::default()).unwrap(); + let db_bench = CanopydbBenchDatabase::new(&db); + benchmark(db_bench, tmpdir.path()) + }; + fs::remove_dir_all(&tmpdir).unwrap(); let mut rows = Vec::new(); @@ -448,6 +444,7 @@ fn main() { sled_results, sanakirja_results, fjall_results, + canopydb_results, ]; let mut identified_smallests = vec![vec![false; results.len()]; rows.len()]; @@ -478,7 +475,16 @@ fn main() { let mut table = comfy_table::Table::new(); table.load_preset(comfy_table::presets::ASCII_MARKDOWN); table.set_width(100); - table.set_header(["", "redb", "lmdb", "rocksdb", "sled", "sanakirja", "fjall"]); + table.set_header([ + "", + "redb", + "lmdb", + "rocksdb", + "sled", + "sanakirja", + "fjall", + "canopydb", + ]); for row in rows { table.add_row(row); }