Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions Cargo.toml
Original file line number Diff line number Diff line change
Expand Up @@ -25,8 +25,8 @@ lockfree = { version = "0.5.1" }
worktable_codegen = { path = "codegen", version = "0.5.5" }
futures = "0.3.30"
uuid = { version = "1.10.0", features = ["v4"] }
data_bucket = "0.2.3"
# data_bucket = { git = "https://github.com/pathscale/DataBucket", branch = "indexset-version-update" }
#data_bucket = "0.2.3"
data_bucket = { git = "https://github.com/pathscale/DataBucket", branch = "main" }
# data_bucket = { path = "../DataBucket", version = "0.2.2" }
performance_measurement_codegen = { path = "performance_measurement/codegen", version = "0.1.0", optional = true }
performance_measurement = { path = "performance_measurement", version = "0.1.0", optional = true }
Expand Down
2 changes: 1 addition & 1 deletion codegen/src/persist_index/generator.rs
Original file line number Diff line number Diff line change
Expand Up @@ -393,7 +393,7 @@ impl Generator {
};
quote! {
let size = get_index_page_size_from_data_length::<#ty>(#const_name);
let #i: #t<_, Link> = #t::new();
let #i: #t<_, Link> = #t::with_maximum_node_size(size);
for page in persisted.#i.1 {
#node
}
Expand Down
6 changes: 4 additions & 2 deletions codegen/src/persist_table/generator/space_file/mod.rs
Original file line number Diff line number Diff line change
Expand Up @@ -154,19 +154,21 @@ impl Generator {
let engine_ident = name_generator.get_persistence_engine_ident();
let dir_name = name_generator.get_dir_name();
let const_name = name_generator.get_page_inner_size_const_ident();
let pk_type = name_generator.get_primary_key_type_ident();

let pk_map = if self.attributes.pk_unsized {
let pk_ident = &self.pk_ident;
quote! {
let pk_map = IndexMap::<#pk_ident, Link, UnsizedNode<_>>::new();
let pk_map = IndexMap::<#pk_ident, Link, UnsizedNode<_>>::with_maximum_node_size(#const_name);
for page in self.primary_index.1 {
let node = page.inner.get_node();
pk_map.attach_node(UnsizedNode::from_inner(node, #const_name));
}
}
} else {
quote! {
let pk_map = IndexMap::new();
let size = get_index_page_size_from_data_length::<#pk_type>(#const_name);
let pk_map = IndexMap::with_maximum_node_size(size);
for page in self.primary_index.1 {
let node = page.inner.get_node();
pk_map.attach_node(node);
Expand Down
54 changes: 52 additions & 2 deletions codegen/src/worktable/generator/index.rs
Original file line number Diff line number Diff line change
Expand Up @@ -15,11 +15,13 @@ impl Generator {
} else {
quote! {}
};
let default_impl = self.gen_index_default_impl();

quote! {
#type_def
#impl_def
#cdc_impl_def
#default_impl
}
}

Expand Down Expand Up @@ -63,11 +65,11 @@ impl Generator {

let derive = if self.is_persist {
quote! {
#[derive(Debug, MemStat, Default, PersistIndex)]
#[derive(Debug, MemStat, PersistIndex)]
}
} else {
quote! {
#[derive(Debug, MemStat, Default)]
#[derive(Debug, MemStat)]
}
};

Expand All @@ -79,6 +81,54 @@ impl Generator {
}
}

fn gen_index_default_impl(&self) -> TokenStream {
let name_generator = WorktableNameGenerator::from_table_name(self.name.to_string());
let index_type_ident = name_generator.get_index_type_ident();
let const_name = name_generator.get_page_inner_size_const_ident();

let index_rows = self
.columns
.indexes
.iter()
.map(|(i, idx)| {
let t = self.columns.columns_map.get(i).unwrap();
let t = if is_float(t.to_string().as_str()) {
quote! { OrderedFloat<#t> }
} else {
quote! { #t }
};
let i = &idx.name;

#[allow(clippy::collapsible_else_if)]
if idx.is_unique {
if is_unsized(&t.to_string()) {
quote! {
#i: IndexMap::with_maximum_node_size(#const_name),
}
} else {
quote! {#i: IndexMap::with_maximum_node_size(get_index_page_size_from_data_length::<#t>(#const_name)),}
}
} else {
if is_unsized(&t.to_string()) {
quote! {#i: IndexMultiMap::with_maximum_node_size(#const_name), }
} else {
quote! {#i: IndexMultiMap::with_maximum_node_size(get_index_page_size_from_data_length::<#t>(#const_name)),}
}
}
})
.collect::<Vec<_>>();

quote! {
impl Default for #index_type_ident {
fn default() -> Self {
Self {
#(#index_rows)*
}
}
}
}
}

/// Generates implementation of `TableSecondaryIndex` trait for index.
fn gen_impl_def(&mut self) -> TokenStream {
let name_generator = WorktableNameGenerator::from_table_name(self.name.to_string());
Expand Down
4 changes: 4 additions & 0 deletions codegen/src/worktable/generator/table/impls.rs
Original file line number Diff line number Diff line change
Expand Up @@ -43,12 +43,16 @@ impl Generator {
let engine = name_generator.get_persistence_engine_ident();
let task = name_generator.get_persistence_task_ident();
let dir_name = name_generator.get_dir_name();
let pk_type = name_generator.get_primary_key_type_ident();
let const_name = name_generator.get_page_inner_size_const_ident();

if self.is_persist {
quote! {
pub async fn new(config: PersistenceConfig) -> eyre::Result<Self> {
let mut inner = WorkTable::default();
inner.table_name = #table_name;
let size = get_index_page_size_from_data_length::<#pk_type>(#const_name);
inner.pk_map = IndexMap::with_maximum_node_size(size);
let table_files_path = format!("{}/{}", config.tables_path, #dir_name);
let engine: #engine = PersistenceEngine::from_table_files_path(table_files_path).await?;
core::result::Result::Ok(Self(
Expand Down
56 changes: 56 additions & 0 deletions tests/persistence/sync/string_re_read.rs
Original file line number Diff line number Diff line change
Expand Up @@ -81,3 +81,59 @@ fn test_key() {
}
})
}

#[test]
fn test_big_amount_reread() {
let config = PersistenceConfig::new("tests/data/key_big_amount", "tests/data/key_big_amount");

let runtime = tokio::runtime::Builder::new_multi_thread()
.worker_threads(2)
.enable_io()
.enable_time()
.build()
.unwrap();

runtime.block_on(async {
remove_dir_if_exists("tests/data/key_big_amount".to_string()).await;

{
let table = StringReReadWorkTable::load_from_file(config.clone())
.await
.unwrap();
for i in 0..1000 {
table
.insert(StringReReadRow {
first: format!("first_{}", i % 100),
id: table.get_next_pk().into(),
third: format!("third_{}", i),
second: format!("second_{}", i),
last: format!("_________________________last_____________________{}", i),
})
.unwrap();
}

table.wait_for_ops().await
}
{
let table = StringReReadWorkTable::load_from_file(config.clone())
.await
.unwrap();
table
.insert(StringReReadRow {
first: "first_last".to_string(),
id: table.get_next_pk().into(),
third: "third_last".to_string(),
second: "second_last".to_string(),
last: "_________________________last_____________________".to_string(),
})
.unwrap();
table.wait_for_ops().await
}
{
let table = StringReReadWorkTable::load_from_file(config.clone())
.await
.unwrap();
assert_eq!(table.select_all().execute().unwrap().len(), 1001);
}
})
}
Loading