pinakes: import in parallel; various UI improvements
Signed-off-by: NotAShelf <raf@notashelf.dev> Change-Id: I1eb47cd79cd4145c56af966f6756fe1d6a6a6964
This commit is contained in:
parent
278bcaa4b0
commit
116fe7b059
42 changed files with 4189 additions and 316 deletions
|
|
@ -111,6 +111,8 @@ fn row_to_media_item(row: &Row) -> rusqlite::Result<MediaItem> {
|
|||
.get::<_, Option<String>>("thumbnail_path")?
|
||||
.map(PathBuf::from),
|
||||
custom_fields: HashMap::new(), // loaded separately
|
||||
// file_mtime may not be present in all queries, so handle gracefully
|
||||
file_mtime: row.get::<_, Option<i64>>("file_mtime").unwrap_or(None),
|
||||
created_at: parse_datetime(&created_str),
|
||||
updated_at: parse_datetime(&updated_str),
|
||||
})
|
||||
|
|
@ -312,18 +314,22 @@ fn load_custom_fields_batch(db: &Connection, items: &mut [MediaItem]) -> rusqlit
|
|||
|
||||
/// Translate a `SearchQuery` into components that can be assembled into SQL.
|
||||
///
|
||||
/// Returns `(fts_expr, where_clauses, join_clauses)` where:
|
||||
/// Returns `(fts_expr, like_terms, where_clauses, join_clauses, params)` where:
|
||||
/// - `fts_expr` is an FTS5 MATCH expression (may be empty),
|
||||
/// - `like_terms` are search terms for LIKE fallback matching,
|
||||
/// - `where_clauses` are extra WHERE predicates (e.g. type filters),
|
||||
/// - `join_clauses` are extra JOIN snippets (e.g. tag filters).
|
||||
/// - `params` are bind parameter values corresponding to `?` placeholders in
|
||||
/// where_clauses and join_clauses.
|
||||
fn search_query_to_fts(query: &SearchQuery) -> (String, Vec<String>, Vec<String>, Vec<String>) {
|
||||
fn search_query_to_fts(
|
||||
query: &SearchQuery,
|
||||
) -> (String, Vec<String>, Vec<String>, Vec<String>, Vec<String>) {
|
||||
let mut wheres = Vec::new();
|
||||
let mut joins = Vec::new();
|
||||
let mut params = Vec::new();
|
||||
let fts = build_fts_expr(query, &mut wheres, &mut joins, &mut params);
|
||||
(fts, wheres, joins, params)
|
||||
let mut like_terms = Vec::new();
|
||||
let fts = build_fts_expr(query, &mut wheres, &mut joins, &mut params, &mut like_terms);
|
||||
(fts, like_terms, wheres, joins, params)
|
||||
}
|
||||
|
||||
fn build_fts_expr(
|
||||
|
|
@ -331,21 +337,35 @@ fn build_fts_expr(
|
|||
wheres: &mut Vec<String>,
|
||||
joins: &mut Vec<String>,
|
||||
params: &mut Vec<String>,
|
||||
like_terms: &mut Vec<String>,
|
||||
) -> String {
|
||||
match query {
|
||||
SearchQuery::FullText(text) => {
|
||||
if text.is_empty() {
|
||||
String::new()
|
||||
} else {
|
||||
sanitize_fts_token(text)
|
||||
// Collect term for LIKE fallback matching
|
||||
like_terms.push(text.clone());
|
||||
// Add implicit prefix matching for better partial matches
|
||||
// This allows "mus" to match "music", "musician", etc.
|
||||
let sanitized = sanitize_fts_token(text);
|
||||
// If it's a single word, add prefix matching
|
||||
if !sanitized.contains(' ') && !sanitized.contains('"') {
|
||||
format!("{}*", sanitized)
|
||||
} else {
|
||||
// For phrases, use as-is but also add NEAR for proximity
|
||||
sanitized
|
||||
}
|
||||
}
|
||||
}
|
||||
SearchQuery::Prefix(prefix) => {
|
||||
like_terms.push(prefix.clone());
|
||||
format!("{}*", sanitize_fts_token(prefix))
|
||||
}
|
||||
SearchQuery::Fuzzy(term) => {
|
||||
// FTS5 does not natively support fuzzy; fall back to prefix match
|
||||
// FTS5 does not natively support fuzzy; use prefix match
|
||||
// as a best-effort approximation.
|
||||
like_terms.push(term.clone());
|
||||
format!("{}*", sanitize_fts_token(term))
|
||||
}
|
||||
SearchQuery::FieldMatch { field, value } => {
|
||||
|
|
@ -355,7 +375,7 @@ fn build_fts_expr(
|
|||
format!("{safe_field}:{safe_value}")
|
||||
}
|
||||
SearchQuery::Not(inner) => {
|
||||
let inner_expr = build_fts_expr(inner, wheres, joins, params);
|
||||
let inner_expr = build_fts_expr(inner, wheres, joins, params, like_terms);
|
||||
if inner_expr.is_empty() {
|
||||
String::new()
|
||||
} else {
|
||||
|
|
@ -365,7 +385,7 @@ fn build_fts_expr(
|
|||
SearchQuery::And(terms) => {
|
||||
let parts: Vec<String> = terms
|
||||
.iter()
|
||||
.map(|t| build_fts_expr(t, wheres, joins, params))
|
||||
.map(|t| build_fts_expr(t, wheres, joins, params, like_terms))
|
||||
.filter(|s| !s.is_empty())
|
||||
.collect();
|
||||
parts.join(" ")
|
||||
|
|
@ -373,7 +393,7 @@ fn build_fts_expr(
|
|||
SearchQuery::Or(terms) => {
|
||||
let parts: Vec<String> = terms
|
||||
.iter()
|
||||
.map(|t| build_fts_expr(t, wheres, joins, params))
|
||||
.map(|t| build_fts_expr(t, wheres, joins, params, like_terms))
|
||||
.filter(|s| !s.is_empty())
|
||||
.collect();
|
||||
if parts.len() <= 1 {
|
||||
|
|
@ -399,6 +419,82 @@ fn build_fts_expr(
|
|||
params.push(tag_name.clone());
|
||||
String::new()
|
||||
}
|
||||
SearchQuery::RangeQuery { field, start, end } => {
|
||||
let col = match field.as_str() {
|
||||
"year" => "m.year",
|
||||
"size" | "file_size" => "m.file_size",
|
||||
"duration" => "m.duration_secs",
|
||||
_ => return String::new(), // Unknown field, ignore
|
||||
};
|
||||
match (start, end) {
|
||||
(Some(s), Some(e)) => {
|
||||
wheres.push(format!("{col} >= ? AND {col} <= ?"));
|
||||
params.push(s.to_string());
|
||||
params.push(e.to_string());
|
||||
}
|
||||
(Some(s), None) => {
|
||||
wheres.push(format!("{col} >= ?"));
|
||||
params.push(s.to_string());
|
||||
}
|
||||
(None, Some(e)) => {
|
||||
wheres.push(format!("{col} <= ?"));
|
||||
params.push(e.to_string());
|
||||
}
|
||||
(None, None) => {}
|
||||
}
|
||||
String::new()
|
||||
}
|
||||
SearchQuery::CompareQuery { field, op, value } => {
|
||||
let col = match field.as_str() {
|
||||
"year" => "m.year",
|
||||
"size" | "file_size" => "m.file_size",
|
||||
"duration" => "m.duration_secs",
|
||||
_ => return String::new(), // Unknown field, ignore
|
||||
};
|
||||
let op_sql = match op {
|
||||
crate::search::CompareOp::GreaterThan => ">",
|
||||
crate::search::CompareOp::GreaterOrEqual => ">=",
|
||||
crate::search::CompareOp::LessThan => "<",
|
||||
crate::search::CompareOp::LessOrEqual => "<=",
|
||||
};
|
||||
wheres.push(format!("{col} {op_sql} ?"));
|
||||
params.push(value.to_string());
|
||||
String::new()
|
||||
}
|
||||
SearchQuery::DateQuery { field, value } => {
|
||||
let col = match field.as_str() {
|
||||
"created" => "m.created_at",
|
||||
"modified" | "updated" => "m.updated_at",
|
||||
_ => return String::new(),
|
||||
};
|
||||
let sql = date_value_to_sqlite_expr(col, value);
|
||||
if !sql.is_empty() {
|
||||
wheres.push(sql);
|
||||
}
|
||||
String::new()
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Convert a DateValue to a SQLite datetime comparison expression
|
||||
fn date_value_to_sqlite_expr(col: &str, value: &crate::search::DateValue) -> String {
|
||||
use crate::search::DateValue;
|
||||
match value {
|
||||
DateValue::Today => format!("date({col}) = date('now')"),
|
||||
DateValue::Yesterday => format!("date({col}) = date('now', '-1 day')"),
|
||||
DateValue::ThisWeek => format!("{col} >= datetime('now', 'weekday 0', '-7 days')"),
|
||||
DateValue::LastWeek => format!(
|
||||
"{col} >= datetime('now', 'weekday 0', '-14 days') AND {col} < datetime('now', 'weekday 0', '-7 days')"
|
||||
),
|
||||
DateValue::ThisMonth => format!("{col} >= datetime('now', 'start of month')"),
|
||||
DateValue::LastMonth => format!(
|
||||
"{col} >= datetime('now', 'start of month', '-1 month') AND {col} < datetime('now', 'start of month')"
|
||||
),
|
||||
DateValue::ThisYear => format!("{col} >= datetime('now', 'start of year')"),
|
||||
DateValue::LastYear => format!(
|
||||
"{col} >= datetime('now', 'start of year', '-1 year') AND {col} < datetime('now', 'start of year')"
|
||||
),
|
||||
DateValue::DaysAgo(days) => format!("{col} >= datetime('now', '-{days} days')"),
|
||||
}
|
||||
}
|
||||
|
||||
|
|
@ -514,8 +610,8 @@ impl StorageBackend for SqliteBackend {
|
|||
db.execute(
|
||||
"INSERT INTO media_items (id, path, file_name, media_type, content_hash, \
|
||||
file_size, title, artist, album, genre, year, duration_secs, description, \
|
||||
thumbnail_path, created_at, updated_at) \
|
||||
VALUES (?1, ?2, ?3, ?4, ?5, ?6, ?7, ?8, ?9, ?10, ?11, ?12, ?13, ?14, ?15, ?16)",
|
||||
thumbnail_path, file_mtime, created_at, updated_at) \
|
||||
VALUES (?1, ?2, ?3, ?4, ?5, ?6, ?7, ?8, ?9, ?10, ?11, ?12, ?13, ?14, ?15, ?16, ?17)",
|
||||
params![
|
||||
item.id.0.to_string(),
|
||||
item.path.to_string_lossy().as_ref(),
|
||||
|
|
@ -533,6 +629,7 @@ impl StorageBackend for SqliteBackend {
|
|||
item.thumbnail_path
|
||||
.as_ref()
|
||||
.map(|p| p.to_string_lossy().to_string()),
|
||||
item.file_mtime,
|
||||
item.created_at.to_rfc3339(),
|
||||
item.updated_at.to_rfc3339(),
|
||||
],
|
||||
|
|
@ -566,7 +663,7 @@ impl StorageBackend for SqliteBackend {
|
|||
let mut stmt = db.prepare(
|
||||
"SELECT id, path, file_name, media_type, content_hash, file_size, \
|
||||
title, artist, album, genre, year, duration_secs, description, \
|
||||
thumbnail_path, created_at, updated_at FROM media_items WHERE id = ?1",
|
||||
thumbnail_path, file_mtime, created_at, updated_at FROM media_items WHERE id = ?1",
|
||||
)?;
|
||||
let mut item = stmt
|
||||
.query_row(params![id.0.to_string()], row_to_media_item)
|
||||
|
|
@ -593,7 +690,7 @@ impl StorageBackend for SqliteBackend {
|
|||
let mut stmt = db.prepare(
|
||||
"SELECT id, path, file_name, media_type, content_hash, file_size, \
|
||||
title, artist, album, genre, year, duration_secs, description, \
|
||||
thumbnail_path, created_at, updated_at FROM media_items WHERE content_hash = ?1",
|
||||
thumbnail_path, file_mtime, created_at, updated_at FROM media_items WHERE content_hash = ?1",
|
||||
)?;
|
||||
let result = stmt
|
||||
.query_row(params![hash.0], row_to_media_item)
|
||||
|
|
@ -609,6 +706,32 @@ impl StorageBackend for SqliteBackend {
|
|||
.map_err(|e| PinakesError::Database(e.to_string()))?
|
||||
}
|
||||
|
||||
async fn get_media_by_path(&self, path: &std::path::Path) -> Result<Option<MediaItem>> {
|
||||
let path_str = path.to_string_lossy().to_string();
|
||||
let conn = Arc::clone(&self.conn);
|
||||
tokio::task::spawn_blocking(move || {
|
||||
let db = conn
|
||||
.lock()
|
||||
.map_err(|e| PinakesError::Database(e.to_string()))?;
|
||||
let mut stmt = db.prepare(
|
||||
"SELECT id, path, file_name, media_type, content_hash, file_size, \
|
||||
title, artist, album, genre, year, duration_secs, description, \
|
||||
thumbnail_path, file_mtime, created_at, updated_at FROM media_items WHERE path = ?1",
|
||||
)?;
|
||||
let result = stmt
|
||||
.query_row(params![path_str], row_to_media_item)
|
||||
.optional()?;
|
||||
if let Some(mut item) = result {
|
||||
item.custom_fields = load_custom_fields_sync(&db, item.id)?;
|
||||
Ok(Some(item))
|
||||
} else {
|
||||
Ok(None)
|
||||
}
|
||||
})
|
||||
.await
|
||||
.map_err(|e| PinakesError::Database(e.to_string()))?
|
||||
}
|
||||
|
||||
async fn list_media(&self, pagination: &Pagination) -> Result<Vec<MediaItem>> {
|
||||
let pagination = pagination.clone();
|
||||
let conn = Arc::clone(&self.conn);
|
||||
|
|
@ -630,7 +753,7 @@ impl StorageBackend for SqliteBackend {
|
|||
let sql = format!(
|
||||
"SELECT id, path, file_name, media_type, content_hash, file_size, \
|
||||
title, artist, album, genre, year, duration_secs, description, \
|
||||
thumbnail_path, created_at, updated_at FROM media_items \
|
||||
thumbnail_path, file_mtime, created_at, updated_at FROM media_items \
|
||||
ORDER BY {order_by} LIMIT ?1 OFFSET ?2"
|
||||
);
|
||||
let mut stmt = db.prepare(&sql)?;
|
||||
|
|
@ -658,7 +781,7 @@ impl StorageBackend for SqliteBackend {
|
|||
"UPDATE media_items SET path = ?2, file_name = ?3, media_type = ?4, \
|
||||
content_hash = ?5, file_size = ?6, title = ?7, artist = ?8, album = ?9, \
|
||||
genre = ?10, year = ?11, duration_secs = ?12, description = ?13, \
|
||||
thumbnail_path = ?14, updated_at = ?15 WHERE id = ?1",
|
||||
thumbnail_path = ?14, file_mtime = ?15, updated_at = ?16 WHERE id = ?1",
|
||||
params![
|
||||
item.id.0.to_string(),
|
||||
item.path.to_string_lossy().as_ref(),
|
||||
|
|
@ -676,6 +799,7 @@ impl StorageBackend for SqliteBackend {
|
|||
item.thumbnail_path
|
||||
.as_ref()
|
||||
.map(|p| p.to_string_lossy().to_string()),
|
||||
item.file_mtime,
|
||||
item.updated_at.to_rfc3339(),
|
||||
],
|
||||
)?;
|
||||
|
|
@ -1067,7 +1191,7 @@ impl StorageBackend for SqliteBackend {
|
|||
.lock()
|
||||
.map_err(|e| PinakesError::Database(e.to_string()))?;
|
||||
|
||||
let (fts_expr, where_clauses, join_clauses, bind_params) =
|
||||
let (fts_expr, _like_terms, where_clauses, join_clauses, bind_params) =
|
||||
search_query_to_fts(&request.query);
|
||||
|
||||
let use_fts = !fts_expr.is_empty();
|
||||
|
|
@ -1309,16 +1433,30 @@ impl StorageBackend for SqliteBackend {
|
|||
}
|
||||
|
||||
async fn batch_delete_media(&self, ids: &[MediaId]) -> Result<u64> {
|
||||
if ids.is_empty() {
|
||||
return Ok(0);
|
||||
}
|
||||
let ids: Vec<String> = ids.iter().map(|id| id.0.to_string()).collect();
|
||||
let conn = Arc::clone(&self.conn);
|
||||
tokio::task::spawn_blocking(move || {
|
||||
let db = conn
|
||||
.lock()
|
||||
.map_err(|e| PinakesError::Database(e.to_string()))?;
|
||||
// Use IN clause for batch delete - much faster than individual deletes
|
||||
// SQLite has a limit of ~500-1000 items in IN clause, so chunk if needed
|
||||
const CHUNK_SIZE: usize = 500;
|
||||
db.execute_batch("BEGIN IMMEDIATE")?;
|
||||
let mut count = 0u64;
|
||||
for id in &ids {
|
||||
let rows = db.execute("DELETE FROM media_items WHERE id = ?1", params![id])?;
|
||||
for chunk in ids.chunks(CHUNK_SIZE) {
|
||||
let placeholders: Vec<String> =
|
||||
(1..=chunk.len()).map(|i| format!("?{}", i)).collect();
|
||||
let sql = format!(
|
||||
"DELETE FROM media_items WHERE id IN ({})",
|
||||
placeholders.join(", ")
|
||||
);
|
||||
let params: Vec<&dyn rusqlite::ToSql> =
|
||||
chunk.iter().map(|s| s as &dyn rusqlite::ToSql).collect();
|
||||
let rows = db.execute(&sql, params.as_slice())?;
|
||||
count += rows as u64;
|
||||
}
|
||||
db.execute_batch("COMMIT")?;
|
||||
|
|
@ -1329,6 +1467,9 @@ impl StorageBackend for SqliteBackend {
|
|||
}
|
||||
|
||||
async fn batch_tag_media(&self, media_ids: &[MediaId], tag_ids: &[Uuid]) -> Result<u64> {
|
||||
if media_ids.is_empty() || tag_ids.is_empty() {
|
||||
return Ok(0);
|
||||
}
|
||||
let media_ids: Vec<String> = media_ids.iter().map(|id| id.0.to_string()).collect();
|
||||
let tag_ids: Vec<String> = tag_ids.iter().map(|id| id.to_string()).collect();
|
||||
let conn = Arc::clone(&self.conn);
|
||||
|
|
@ -1337,13 +1478,14 @@ impl StorageBackend for SqliteBackend {
|
|||
.lock()
|
||||
.map_err(|e| PinakesError::Database(e.to_string()))?;
|
||||
db.execute_batch("BEGIN IMMEDIATE")?;
|
||||
// Prepare statement once for reuse
|
||||
let mut stmt = db.prepare_cached(
|
||||
"INSERT OR IGNORE INTO media_tags (media_id, tag_id) VALUES (?1, ?2)",
|
||||
)?;
|
||||
let mut count = 0u64;
|
||||
for mid in &media_ids {
|
||||
for tid in &tag_ids {
|
||||
db.execute(
|
||||
"INSERT OR IGNORE INTO media_tags (media_id, tag_id) VALUES (?1, ?2)",
|
||||
params![mid, tid],
|
||||
)?;
|
||||
stmt.execute(params![mid, tid])?;
|
||||
count += 1;
|
||||
}
|
||||
}
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue