Improve RocksDB and SpeeDB performance (#3000)

This commit is contained in:
Tobie Morgan Hitchcock 2023-11-20 18:47:44 +00:00 committed by GitHub
parent 6efd3e3d87
commit 9c0d107588
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
13 changed files with 227 additions and 172 deletions

1
Cargo.lock generated
View file

@ -5044,6 +5044,7 @@ dependencies = [
"nanoid", "nanoid",
"native-tls", "native-tls",
"nom", "nom",
"num_cpus",
"once_cell", "once_cell",
"path-clean", "path-clean",
"pbkdf2", "pbkdf2",

View file

@ -86,6 +86,7 @@ md-5 = "0.10.6"
nanoid = "0.4.0" nanoid = "0.4.0"
native-tls = { version = "0.2.11", optional = true } native-tls = { version = "0.2.11", optional = true }
nom = { version = "7.1.3", features = ["alloc"] } nom = { version = "7.1.3", features = ["alloc"] }
num_cpus = "1.16.0"
once_cell = "1.18.0" once_cell = "1.18.0"
path-clean = "1.0.1" path-clean = "1.0.1"
pbkdf2 = { version = "0.12.2", features = ["simple"] } pbkdf2 = { version = "0.12.2", features = ["simple"] }
@ -96,7 +97,7 @@ regex = "1.10.2"
reqwest = { version = "0.11.22", default-features = false, features = ["json", "stream", "multipart"], optional = true } reqwest = { version = "0.11.22", default-features = false, features = ["json", "stream", "multipart"], optional = true }
revision = "0.5.0" revision = "0.5.0"
roaring = { version = "0.10.2", features = ["serde"] } roaring = { version = "0.10.2", features = ["serde"] }
rocksdb = { version = "0.21.0", optional = true } rocksdb = { version = "0.21.0", features = ["lz4", "snappy"], optional = true }
rust_decimal = { version = "1.32.0", features = ["maths", "serde-str"] } rust_decimal = { version = "1.32.0", features = ["maths", "serde-str"] }
rust-stemmers = "1.2.0" rust-stemmers = "1.2.0"
rustls = { version = "=0.21.6", optional = true } rustls = { version = "=0.21.6", optional = true }
@ -107,7 +108,7 @@ serde_json = "1.0.108"
sha1 = "0.10.6" sha1 = "0.10.6"
sha2 = "0.10.8" sha2 = "0.10.8"
snap = "1.1.0" snap = "1.1.0"
speedb = { version = "0.0.2", optional = true } speedb = { version = "0.0.2", features = ["lz4", "snappy"], optional = true }
storekey = "0.5.0" storekey = "0.5.0"
thiserror = "1.0.50" thiserror = "1.0.50"
tikv = { version = "0.2.0-surreal.2", default-features = false, package = "surrealdb-tikv-client", optional = true } tikv = { version = "0.2.0-surreal.2", default-features = false, package = "surrealdb-tikv-client", optional = true }

View file

@ -110,7 +110,7 @@ impl<'a> Document<'a> {
} }
} }
// Remove metadata fields on output // Remove metadata fields on output
out.del(ctx, opt, txn, &*META).await?; out.cut(&*META);
// Output result // Output result
Ok(out) Ok(out)
} }

View file

@ -1,75 +0,0 @@
//! Stores a DEFINE LOGIN ON DATABASE config definition
use crate::key::error::KeyCategory;
use crate::key::key_req::KeyRequirements;
use derive::Key;
use serde::{Deserialize, Serialize};
#[derive(Clone, Debug, Eq, PartialEq, PartialOrd, Serialize, Deserialize, Key)]
pub struct Lg<'a> {
__: u8,
_a: u8,
pub ns: &'a str,
_b: u8,
pub db: &'a str,
_c: u8,
_d: u8,
_e: u8,
pub dl: &'a str,
}
pub fn new<'a>(ns: &'a str, db: &'a str, dl: &'a str) -> Lg<'a> {
Lg::new(ns, db, dl)
}
pub fn prefix(ns: &str, db: &str) -> Vec<u8> {
let mut k = super::all::new(ns, db).encode().unwrap();
k.extend_from_slice(&[b'!', b'l', b'g', 0x00]);
k
}
pub fn suffix(ns: &str, db: &str) -> Vec<u8> {
let mut k = super::all::new(ns, db).encode().unwrap();
k.extend_from_slice(&[b'!', b'l', b'g', 0xff]);
k
}
impl KeyRequirements for Lg<'_> {
fn key_category(&self) -> KeyCategory {
KeyCategory::DatabaseLog
}
}
impl<'a> Lg<'a> {
pub fn new(ns: &'a str, db: &'a str, dl: &'a str) -> Self {
Self {
__: b'/',
_a: b'*',
ns,
_b: b'*',
db,
_c: b'!',
_d: b'l',
_e: b'g',
dl,
}
}
}
#[cfg(test)]
mod tests {
#[test]
fn key() {
use super::*;
#[rustfmt::skip]
let val = Lg::new(
"testns",
"testdb",
"testdl",
);
let enc = Lg::encode(&val).unwrap();
assert_eq!(enc, b"/*testns\0*testdb\0!lgtestdl\0");
let dec = Lg::decode(&enc).unwrap();
assert_eq!(val, dec);
}
}

View file

@ -1,7 +1,6 @@
pub mod all; pub mod all;
pub mod az; pub mod az;
pub mod fc; pub mod fc;
pub mod lg;
pub mod pa; pub mod pa;
pub mod sc; pub mod sc;
pub mod tb; pub mod tb;

View file

@ -6,8 +6,7 @@ use crate::kvs::Key;
/// sprint_key converts a key to an escaped string. /// sprint_key converts a key to an escaped string.
/// This is used for logging and debugging tests and should not be used in implementation code. /// This is used for logging and debugging tests and should not be used in implementation code.
pub fn sprint_key(key: &Key) -> String { pub fn sprint_key(key: &Key) -> String {
key.clone() key.iter()
.iter()
.flat_map(|&byte| std::ascii::escape_default(byte)) .flat_map(|&byte| std::ascii::escape_default(byte))
.map(|byte| byte as char) .map(|byte| byte as char)
.collect::<String>() .collect::<String>()

View file

@ -1,70 +0,0 @@
//! Stores a DEFINE LOGIN ON NAMESPACE config definition
use crate::key::error::KeyCategory;
use crate::key::key_req::KeyRequirements;
use derive::Key;
use serde::{Deserialize, Serialize};
#[derive(Clone, Debug, Eq, PartialEq, PartialOrd, Serialize, Deserialize, Key)]
pub struct Lg<'a> {
__: u8,
_a: u8,
pub ns: &'a str,
_b: u8,
_c: u8,
_d: u8,
pub us: &'a str,
}
pub fn new<'a>(ns: &'a str, us: &'a str) -> Lg<'a> {
Lg::new(ns, us)
}
pub fn prefix(ns: &str) -> Vec<u8> {
let mut k = super::all::new(ns).encode().unwrap();
k.extend_from_slice(&[b'!', b'l', b'g', 0x00]);
k
}
pub fn suffix(ns: &str) -> Vec<u8> {
let mut k = super::all::new(ns).encode().unwrap();
k.extend_from_slice(&[b'!', b'l', b'g', 0xff]);
k
}
impl KeyRequirements for Lg<'_> {
fn key_category(&self) -> KeyCategory {
KeyCategory::DatabaseLogAlias
}
}
impl<'a> Lg<'a> {
pub fn new(ns: &'a str, us: &'a str) -> Self {
Self {
__: b'/',
_a: b'*',
ns,
_b: b'!',
_c: b'l',
_d: b'g',
us,
}
}
}
#[cfg(test)]
mod tests {
#[test]
fn key() {
use super::*;
#[rustfmt::skip]
let val = Lg::new(
"testns",
"testus",
);
let enc = Lg::encode(&val).unwrap();
assert_eq!(enc, b"/*testns\0!lgtestus\0");
let dec = Lg::decode(&enc).unwrap();
assert_eq!(val, dec);
}
}

View file

@ -1,6 +1,5 @@
pub mod all; pub mod all;
pub mod db; pub mod db;
pub mod di; pub mod di;
pub mod lg;
pub mod tk; pub mod tk;
pub mod us; pub mod us;

View file

@ -0,0 +1,57 @@
use once_cell::sync::Lazy;
pub static ROCKSDB_THREAD_COUNT: Lazy<i32> = Lazy::new(|| {
let default = num_cpus::get() as i32;
std::env::var("SURREAL_ROCKSDB_THREAD_COUNT")
.map(|v| v.parse::<i32>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_WRITE_BUFFER_SIZE: Lazy<usize> = Lazy::new(|| {
let default = 256 * 1024 * 1024;
std::env::var("SURREAL_ROCKSDB_WRITE_BUFFER_SIZE")
.map(|v| v.parse::<usize>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_TARGET_FILE_SIZE_BASE: Lazy<u64> = Lazy::new(|| {
let default = 512 * 1024 * 1024;
std::env::var("SURREAL_ROCKSDB_TARGET_FILE_SIZE_BASE")
.map(|v| v.parse::<u64>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_MAX_WRITE_BUFFER_NUMBER: Lazy<i32> = Lazy::new(|| {
let default = 32;
std::env::var("SURREAL_ROCKSDB_MAX_WRITE_BUFFER_NUMBER")
.map(|v| v.parse::<i32>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_MIN_WRITE_BUFFER_NUMBER_TO_MERGE: Lazy<i32> = Lazy::new(|| {
let default = 4;
std::env::var("SURREAL_ROCKSDB_MIN_WRITE_BUFFER_NUMBER_TO_MERGE")
.map(|v| v.parse::<i32>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_ENABLE_PIPELINED_WRITES: Lazy<bool> = Lazy::new(|| {
let default = true;
std::env::var("SURREAL_ROCKSDB_ENABLE_PIPELINED_WRITES")
.map(|v| v.parse::<bool>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_ENABLE_BLOB_FILES: Lazy<bool> = Lazy::new(|| {
let default = true;
std::env::var("SURREAL_ROCKSDB_ENABLE_BLOB_FILES")
.map(|v| v.parse::<bool>().unwrap_or(default))
.unwrap_or(default)
});
pub static ROCKSDB_MIN_BLOB_SIZE: Lazy<u64> = Lazy::new(|| {
let default = 4 * 1024;
std::env::var("SURREAL_ROCKSDB_MIN_BLOB_SIZE")
.map(|v| v.parse::<u64>().unwrap_or(default))
.unwrap_or(default)
});

View file

@ -1,5 +1,7 @@
#![cfg(feature = "kv-rocksdb")] #![cfg(feature = "kv-rocksdb")]
mod cnf;
use crate::err::Error; use crate::err::Error;
use crate::key::error::KeyCategory; use crate::key::error::KeyCategory;
use crate::kvs::Check; use crate::kvs::Check;
@ -7,7 +9,10 @@ use crate::kvs::Key;
use crate::kvs::Val; use crate::kvs::Val;
use crate::vs::{try_to_u64_be, u64_to_versionstamp, Versionstamp}; use crate::vs::{try_to_u64_be, u64_to_versionstamp, Versionstamp};
use futures::lock::Mutex; use futures::lock::Mutex;
use rocksdb::{OptimisticTransactionDB, OptimisticTransactionOptions, ReadOptions, WriteOptions}; use rocksdb::{
DBCompactionStyle, DBCompressionType, LogLevel, OptimisticTransactionDB,
OptimisticTransactionOptions, Options, ReadOptions, WriteOptions,
};
use std::ops::Range; use std::ops::Range;
use std::pin::Pin; use std::pin::Pin;
use std::sync::Arc; use std::sync::Arc;
@ -68,17 +73,57 @@ impl Drop for Transaction {
impl Datastore { impl Datastore {
/// Open a new database /// Open a new database
pub(crate) async fn new(path: &str) -> Result<Datastore, Error> { pub(crate) async fn new(path: &str) -> Result<Datastore, Error> {
// Configure custom options
let mut opts = Options::default();
// Ensure we use fdatasync
opts.set_use_fsync(false);
// Only use warning log level
opts.set_log_level(LogLevel::Warn);
// Create database if missing
opts.create_if_missing(true);
// Create column families if missing
opts.create_missing_column_families(true);
// Set the datastore compaction style
opts.set_compaction_style(DBCompactionStyle::Level);
// Increase the background thread count
opts.increase_parallelism(*cnf::ROCKSDB_THREAD_COUNT);
// Set the maximum number of write buffers
opts.set_max_write_buffer_number(*cnf::ROCKSDB_MAX_WRITE_BUFFER_NUMBER);
// Set the amount of data to build up in memory
opts.set_write_buffer_size(*cnf::ROCKSDB_WRITE_BUFFER_SIZE);
// Set the target file size for compaction
opts.set_target_file_size_base(*cnf::ROCKSDB_TARGET_FILE_SIZE_BASE);
// Set minimum number of write buffers to merge
opts.set_min_write_buffer_number_to_merge(*cnf::ROCKSDB_MIN_WRITE_BUFFER_NUMBER_TO_MERGE);
// Use separate write thread queues
opts.set_enable_pipelined_write(*cnf::ROCKSDB_ENABLE_PIPELINED_WRITES);
// Enable separation of keys and values
opts.set_enable_blob_files(*cnf::ROCKSDB_ENABLE_BLOB_FILES);
// Store 4KB values separate from keys
opts.set_min_blob_size(*cnf::ROCKSDB_MIN_BLOB_SIZE);
// Set specific compression levels
opts.set_compression_per_level(&[
DBCompressionType::None,
DBCompressionType::None,
DBCompressionType::Lz4hc,
DBCompressionType::Lz4hc,
DBCompressionType::Lz4hc,
]);
// Create the datastore
Ok(Datastore { Ok(Datastore {
db: Arc::pin(OptimisticTransactionDB::open_default(path)?), db: Arc::pin(OptimisticTransactionDB::open(&opts, path)?),
}) })
} }
/// Start a new transaction /// Start a new transaction
pub(crate) async fn transaction(&self, write: bool, _: bool) -> Result<Transaction, Error> { pub(crate) async fn transaction(&self, write: bool, _: bool) -> Result<Transaction, Error> {
// Activate the snapshot options // Set the transaction options
let mut to = OptimisticTransactionOptions::default(); let mut to = OptimisticTransactionOptions::default();
to.set_snapshot(true); to.set_snapshot(true);
// Set the write options
let mut wo = WriteOptions::default();
wo.set_sync(false);
// Create a new transaction // Create a new transaction
let inner = self.db.transaction_opt(&WriteOptions::default(), &to); let inner = self.db.transaction_opt(&wo, &to);
// The database reference must always outlive // The database reference must always outlive
// the transaction. If it doesn't then this // the transaction. If it doesn't then this
// is undefined behaviour. This unsafe block // is undefined behaviour. This unsafe block
@ -93,6 +138,8 @@ impl Datastore {
}; };
let mut ro = ReadOptions::default(); let mut ro = ReadOptions::default();
ro.set_snapshot(&inner.snapshot()); ro.set_snapshot(&inner.snapshot());
ro.set_async_io(true);
ro.fill_cache(true);
// Specify the check level // Specify the check level
#[cfg(not(debug_assertions))] #[cfg(not(debug_assertions))]
let check = Check::Warn; let check = Check::Warn;

57
lib/src/kvs/speedb/cnf.rs Normal file
View file

@ -0,0 +1,57 @@
use once_cell::sync::Lazy;
pub static SPEEDB_THREAD_COUNT: Lazy<i32> = Lazy::new(|| {
let default = num_cpus::get() as i32;
std::env::var("SURREAL_SPEEDB_THREAD_COUNT")
.map(|v| v.parse::<i32>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_WRITE_BUFFER_SIZE: Lazy<usize> = Lazy::new(|| {
let default = 256 * 1024 * 1024;
std::env::var("SURREAL_SPEEDB_WRITE_BUFFER_SIZE")
.map(|v| v.parse::<usize>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_TARGET_FILE_SIZE_BASE: Lazy<u64> = Lazy::new(|| {
let default = 512 * 1024 * 1024;
std::env::var("SURREAL_SPEEDB_TARGET_FILE_SIZE_BASE")
.map(|v| v.parse::<u64>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_MAX_WRITE_BUFFER_NUMBER: Lazy<i32> = Lazy::new(|| {
let default = 32;
std::env::var("SURREAL_SPEEDB_MAX_WRITE_BUFFER_NUMBER")
.map(|v| v.parse::<i32>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_MIN_WRITE_BUFFER_NUMBER_TO_MERGE: Lazy<i32> = Lazy::new(|| {
let default = 4;
std::env::var("SURREAL_SPEEDB_MIN_WRITE_BUFFER_NUMBER_TO_MERGE")
.map(|v| v.parse::<i32>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_ENABLE_PIPELINED_WRITES: Lazy<bool> = Lazy::new(|| {
let default = true;
std::env::var("SURREAL_SPEEDB_ENABLE_PIPELINED_WRITES")
.map(|v| v.parse::<bool>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_ENABLE_BLOB_FILES: Lazy<bool> = Lazy::new(|| {
let default = true;
std::env::var("SURREAL_SPEEDB_ENABLE_BLOB_FILES")
.map(|v| v.parse::<bool>().unwrap_or(default))
.unwrap_or(default)
});
pub static SPEEDB_MIN_BLOB_SIZE: Lazy<u64> = Lazy::new(|| {
let default = 4 * 1024;
std::env::var("SURREAL_SPEEDB_ENABLE_BLOB_FILES")
.map(|v| v.parse::<u64>().unwrap_or(default))
.unwrap_or(default)
});

View file

@ -1,5 +1,7 @@
#![cfg(feature = "kv-speedb")] #![cfg(feature = "kv-speedb")]
mod cnf;
use crate::err::Error; use crate::err::Error;
use crate::key::error::KeyCategory; use crate::key::error::KeyCategory;
use crate::kvs::Check; use crate::kvs::Check;
@ -7,7 +9,10 @@ use crate::kvs::Key;
use crate::kvs::Val; use crate::kvs::Val;
use crate::vs::{try_to_u64_be, u64_to_versionstamp, Versionstamp}; use crate::vs::{try_to_u64_be, u64_to_versionstamp, Versionstamp};
use futures::lock::Mutex; use futures::lock::Mutex;
use speedb::{OptimisticTransactionDB, OptimisticTransactionOptions, ReadOptions, WriteOptions}; use speedb::{
DBCompactionStyle, DBCompressionType, LogLevel, OptimisticTransactionDB,
OptimisticTransactionOptions, Options, ReadOptions, WriteOptions,
};
use std::ops::Range; use std::ops::Range;
use std::pin::Pin; use std::pin::Pin;
use std::sync::Arc; use std::sync::Arc;
@ -68,17 +73,57 @@ impl Drop for Transaction {
impl Datastore { impl Datastore {
/// Open a new database /// Open a new database
pub(crate) async fn new(path: &str) -> Result<Datastore, Error> { pub(crate) async fn new(path: &str) -> Result<Datastore, Error> {
// Configure custom options
let mut opts = Options::default();
// Ensure we use fdatasync
opts.set_use_fsync(false);
// Only use warning log level
opts.set_log_level(LogLevel::Warn);
// Create database if missing
opts.create_if_missing(true);
// Create column families if missing
opts.create_missing_column_families(true);
// Set the datastore compaction style
opts.set_compaction_style(DBCompactionStyle::Level);
// Increase the background thread count
opts.increase_parallelism(*cnf::SPEEDB_THREAD_COUNT);
// Set the maximum number of write buffers
opts.set_max_write_buffer_number(*cnf::SPEEDB_MAX_WRITE_BUFFER_NUMBER);
// Set the amount of data to build up in memory
opts.set_write_buffer_size(*cnf::SPEEDB_WRITE_BUFFER_SIZE);
// Set the target file size for compaction
opts.set_target_file_size_base(*cnf::SPEEDB_TARGET_FILE_SIZE_BASE);
// Set minimum number of write buffers to merge
opts.set_min_write_buffer_number_to_merge(*cnf::SPEEDB_MIN_WRITE_BUFFER_NUMBER_TO_MERGE);
// Use separate write thread queues
opts.set_enable_pipelined_write(*cnf::SPEEDB_ENABLE_PIPELINED_WRITES);
// Enable separation of keys and values
opts.set_enable_blob_files(*cnf::SPEEDB_ENABLE_BLOB_FILES);
// Store 4KB values separate from keys
opts.set_min_blob_size(*cnf::SPEEDB_MIN_BLOB_SIZE);
// Set specific compression levels
opts.set_compression_per_level(&[
DBCompressionType::None,
DBCompressionType::None,
DBCompressionType::Lz4hc,
DBCompressionType::Lz4hc,
DBCompressionType::Lz4hc,
]);
// Create the datastore
Ok(Datastore { Ok(Datastore {
db: Arc::pin(OptimisticTransactionDB::open_default(path)?), db: Arc::pin(OptimisticTransactionDB::open(&opts, path)?),
}) })
} }
/// Start a new transaction /// Start a new transaction
pub(crate) async fn transaction(&self, write: bool, _: bool) -> Result<Transaction, Error> { pub(crate) async fn transaction(&self, write: bool, _: bool) -> Result<Transaction, Error> {
// Activate the snapshot options // Set the transaction options
let mut to = OptimisticTransactionOptions::default(); let mut to = OptimisticTransactionOptions::default();
to.set_snapshot(true); to.set_snapshot(true);
// Set the write options
let mut wo = WriteOptions::default();
wo.set_sync(false);
// Create a new transaction // Create a new transaction
let inner = self.db.transaction_opt(&WriteOptions::default(), &to); let inner = self.db.transaction_opt(&wo, &to);
// The database reference must always outlive // The database reference must always outlive
// the transaction. If it doesn't then this // the transaction. If it doesn't then this
// is undefined behaviour. This unsafe block // is undefined behaviour. This unsafe block
@ -93,6 +138,7 @@ impl Datastore {
}; };
let mut ro = ReadOptions::default(); let mut ro = ReadOptions::default();
ro.set_snapshot(&inner.snapshot()); ro.set_snapshot(&inner.snapshot());
ro.fill_cache(true);
// Specify the check level // Specify the check level
#[cfg(not(debug_assertions))] #[cfg(not(debug_assertions))]
let check = Check::Warn; let check = Check::Warn;

View file

@ -8,8 +8,6 @@ use crate::dbs::node::Timestamp;
use crate::err::Error; use crate::err::Error;
use crate::idg::u32::U32; use crate::idg::u32::U32;
use crate::idx::trees::store::TreeStoreType; use crate::idx::trees::store::TreeStoreType;
#[cfg(debug_assertions)]
use crate::key::debug;
use crate::key::error::KeyCategory; use crate::key::error::KeyCategory;
use crate::key::key_req::KeyRequirements; use crate::key::key_req::KeyRequirements;
use crate::kvs::cache::Cache; use crate::kvs::cache::Cache;
@ -290,10 +288,10 @@ impl Transaction {
#[allow(unused_variables)] #[allow(unused_variables)]
pub async fn del<K>(&mut self, key: K) -> Result<(), Error> pub async fn del<K>(&mut self, key: K) -> Result<(), Error>
where where
K: Into<Key> + Debug + Into<Vec<u8>> + Clone, K: Into<Key> + Debug,
{ {
#[cfg(debug_assertions)] #[cfg(debug_assertions)]
trace!("Del {:?}", crate::key::debug::sprint_key(&key.clone().into())); trace!("Del {:?}", key);
match self { match self {
#[cfg(feature = "kv-mem")] #[cfg(feature = "kv-mem")]
Transaction { Transaction {
@ -544,7 +542,7 @@ impl Transaction {
val: V, val: V,
) -> Result<(), Error> ) -> Result<(), Error>
where where
K: Into<Key> + Debug + Clone, K: Into<Key> + Debug,
V: Into<Val> + Debug, V: Into<Val> + Debug,
{ {
#[cfg(debug_assertions)] #[cfg(debug_assertions)]
@ -651,14 +649,10 @@ impl Transaction {
#[allow(unused_variables)] #[allow(unused_variables)]
pub async fn scan<K>(&mut self, rng: Range<K>, limit: u32) -> Result<Vec<(Key, Val)>, Error> pub async fn scan<K>(&mut self, rng: Range<K>, limit: u32) -> Result<Vec<(Key, Val)>, Error>
where where
K: Into<Key> + Debug + Clone, K: Into<Key> + Debug,
{ {
#[cfg(debug_assertions)] #[cfg(debug_assertions)]
trace!( trace!("Scan {:?} - {:?}", rng.start, rng.end);
"Scan {:?} - {:?}",
debug::sprint_key(&rng.start.clone().into()),
debug::sprint_key(&rng.end.clone().into())
);
match self { match self {
#[cfg(feature = "kv-mem")] #[cfg(feature = "kv-mem")]
Transaction { Transaction {
@ -794,7 +788,7 @@ impl Transaction {
/// This function fetches key-value pairs from the underlying datastore in batches of 1000. /// This function fetches key-value pairs from the underlying datastore in batches of 1000.
pub async fn getr<K>(&mut self, rng: Range<K>, limit: u32) -> Result<Vec<(Key, Val)>, Error> pub async fn getr<K>(&mut self, rng: Range<K>, limit: u32) -> Result<Vec<(Key, Val)>, Error>
where where
K: Into<Key> + Debug + Clone, K: Into<Key> + Debug,
{ {
#[cfg(debug_assertions)] #[cfg(debug_assertions)]
trace!("Getr {:?}..{:?} (limit: {limit})", rng.start, rng.end); trace!("Getr {:?}..{:?} (limit: {limit})", rng.start, rng.end);