From 619900e4d38a280b9e961418ef0d9e14f391f267 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Cl=C3=A9ment=20Renault?= Date: Tue, 12 Aug 2025 16:20:32 +0200 Subject: [PATCH] Make clippy happy --- crates/milli/src/progress.rs | 1 - .../milli/src/update/index_documents/mod.rs | 2 +- crates/milli/src/vector/mod.rs | 102 ++++++++---------- 3 files changed, 46 insertions(+), 59 deletions(-) diff --git a/crates/milli/src/progress.rs b/crates/milli/src/progress.rs index 728e95652..eb309b0b0 100644 --- a/crates/milli/src/progress.rs +++ b/crates/milli/src/progress.rs @@ -5,7 +5,6 @@ use std::sync::atomic::{AtomicU32, AtomicUsize, Ordering}; use std::sync::{Arc, RwLock}; use std::time::{Duration, Instant}; -use enum_iterator::Sequence; use indexmap::IndexMap; use itertools::Itertools; use serde::Serialize; diff --git a/crates/milli/src/update/index_documents/mod.rs b/crates/milli/src/update/index_documents/mod.rs index 022146e88..465a40f42 100644 --- a/crates/milli/src/update/index_documents/mod.rs +++ b/crates/milli/src/update/index_documents/mod.rs @@ -485,7 +485,7 @@ where // If an embedder wasn't used in the typedchunk but must be binary quantized // we should insert it in `dimension` - let index_version = self.index.get_version(&self.wtxn)?.unwrap(); + let index_version = self.index.get_version(self.wtxn)?.unwrap(); for (name, action) in settings_diff.embedding_config_updates.iter() { if action.is_being_quantized && !dimension.contains_key(name.as_str()) { let index = self.index.embedding_configs().embedder_id(self.wtxn, name)?.ok_or( diff --git a/crates/milli/src/vector/mod.rs b/crates/milli/src/vector/mod.rs index 8b3ba7577..bf4ae7ba4 100644 --- a/crates/milli/src/vector/mod.rs +++ b/crates/milli/src/vector/mod.rs @@ -129,14 +129,12 @@ impl VectorStore { self._arroy_items_in_store(rtxn, self.arroy_angular_db(), store_id, with_items) .map_err(Into::into) } + } else if self.quantized { + self._items_in_store(rtxn, self.quantized_db(), store_id, with_items) + .map_err(Into::into) } else { - if self.quantized { - self._items_in_store(rtxn, self.quantized_db(), store_id, with_items) - .map_err(Into::into) - } else { - self._items_in_store(rtxn, self.angular_db(), store_id, with_items) - .map_err(Into::into) - } + self._items_in_store(rtxn, self.angular_db(), store_id, with_items) + .map_err(Into::into) } } @@ -193,20 +191,18 @@ impl VectorStore { .transpose()? .map(|reader| reader.dimensions())) } + } else if self.quantized { + Ok(self + .readers(rtxn, self.quantized_db()) + .next() + .transpose()? + .map(|reader| reader.dimensions())) } else { - if self.quantized { - Ok(self - .readers(rtxn, self.quantized_db()) - .next() - .transpose()? - .map(|reader| reader.dimensions())) - } else { - Ok(self - .readers(rtxn, self.angular_db()) - .next() - .transpose()? - .map(|reader| reader.dimensions())) - } + Ok(self + .readers(rtxn, self.angular_db()) + .next() + .transpose()? + .map(|reader| reader.dimensions())) } } @@ -550,20 +546,18 @@ impl VectorStore { } writer.contains_item(rtxn, item)? } - } else { - if self.quantized { - let writer = hannoy::Writer::new(self.quantized_db(), index, dimension); - if writer.is_empty(rtxn)? { - continue; - } - writer.contains_item(rtxn, item)? - } else { - let writer = hannoy::Writer::new(self.angular_db(), index, dimension); - if writer.is_empty(rtxn)? { - continue; - } - writer.contains_item(rtxn, item)? + } else if self.quantized { + let writer = hannoy::Writer::new(self.quantized_db(), index, dimension); + if writer.is_empty(rtxn)? { + continue; } + writer.contains_item(rtxn, item)? + } else { + let writer = hannoy::Writer::new(self.angular_db(), index, dimension); + if writer.is_empty(rtxn)? { + continue; + } + writer.contains_item(rtxn, item)? }; if contains { return Ok(contains); @@ -587,13 +581,11 @@ impl VectorStore { self._arroy_nns_by_item(rtxn, self.arroy_angular_db(), item, limit, filter) .map_err(Into::into) } + } else if self.quantized { + self._nns_by_item(rtxn, self.quantized_db(), item, limit, filter) + .map_err(Into::into) } else { - if self.quantized { - self._nns_by_item(rtxn, self.quantized_db(), item, limit, filter) - .map_err(Into::into) - } else { - self._nns_by_item(rtxn, self.angular_db(), item, limit, filter).map_err(Into::into) - } + self._nns_by_item(rtxn, self.angular_db(), item, limit, filter).map_err(Into::into) } } @@ -669,14 +661,12 @@ impl VectorStore { self._arroy_nns_by_vector(rtxn, self.arroy_angular_db(), vector, limit, filter) .map_err(Into::into) } + } else if self.quantized { + self._nns_by_vector(rtxn, self.quantized_db(), vector, limit, filter) + .map_err(Into::into) } else { - if self.quantized { - self._nns_by_vector(rtxn, self.quantized_db(), vector, limit, filter) - .map_err(Into::into) - } else { - self._nns_by_vector(rtxn, self.angular_db(), vector, limit, filter) - .map_err(Into::into) - } + self._nns_by_vector(rtxn, self.angular_db(), vector, limit, filter) + .map_err(Into::into) } } @@ -754,18 +744,16 @@ impl VectorStore { } } } - } else { - if self.quantized { - for reader in self.readers(rtxn, self.quantized_db()) { - if let Some(vec) = reader?.item_vector(rtxn, item_id)? { - vectors.push(vec); - } + } else if self.quantized { + for reader in self.readers(rtxn, self.quantized_db()) { + if let Some(vec) = reader?.item_vector(rtxn, item_id)? { + vectors.push(vec); } - } else { - for reader in self.readers(rtxn, self.angular_db()) { - if let Some(vec) = reader?.item_vector(rtxn, item_id)? { - vectors.push(vec); - } + } + } else { + for reader in self.readers(rtxn, self.angular_db()) { + if let Some(vec) = reader?.item_vector(rtxn, item_id)? { + vectors.push(vec); } } }