Phase 1 (discovery): walkdir + filename-only metadata, zero archive I/O. Books are visible immediately in the UI while Phase 2 runs in background. Phase 2 (analysis): open each archive once via analyze_book() to extract page_count and first page bytes, then generate WebP thumbnail directly in the indexer — removing the HTTP roundtrip to the API checkup endpoint. - Add parse_metadata_fast() (infallible, no archive I/O) - Add analyze_book() returning (page_count, first_page_bytes) in one pass - Add looks_like_image() magic bytes check for unrar p stdout validation - Add lsar fallback in list_cbr_images() for UTF-16BE encoded filenames - Add directory_mtimes table to skip unchanged dirs on incremental scans - Add analyzer.rs: generate_thumbnail, analyze_library_books, regenerate_thumbnails - Remove run_checkup() from API; indexer handles thumbnail jobs directly - Remove api_base_url/api_bootstrap_token from IndexerConfig and AppState - Add unar + poppler-utils to indexer Dockerfile - Fix smoke.sh: wait for job completion, check thumbnail_url field Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
181 lines
6.7 KiB
Rust
181 lines
6.7 KiB
Rust
use anyhow::{Context, Result};
|
|
use chrono::{DateTime, Utc};
|
|
use reqwest::Client;
|
|
use serde::Serialize;
|
|
use sqlx::{PgPool, Row};
|
|
use tracing::info;
|
|
use uuid::Uuid;
|
|
|
|
#[derive(Serialize)]
|
|
struct SearchDoc {
|
|
id: String,
|
|
library_id: String,
|
|
kind: String,
|
|
title: String,
|
|
author: Option<String>,
|
|
series: Option<String>,
|
|
volume: Option<i32>,
|
|
language: Option<String>,
|
|
}
|
|
|
|
pub async fn sync_meili(pool: &PgPool, meili_url: &str, meili_master_key: &str) -> Result<()> {
|
|
let client = Client::new();
|
|
let base = meili_url.trim_end_matches('/');
|
|
|
|
// Ensure index exists and has proper settings
|
|
let _ = client
|
|
.post(format!("{base}/indexes"))
|
|
.header("Authorization", format!("Bearer {meili_master_key}"))
|
|
.json(&serde_json::json!({"uid": "books", "primaryKey": "id"}))
|
|
.send()
|
|
.await;
|
|
|
|
let _ = client
|
|
.patch(format!("{base}/indexes/books/settings/filterable-attributes"))
|
|
.header("Authorization", format!("Bearer {meili_master_key}"))
|
|
.json(&serde_json::json!(["library_id", "kind"]))
|
|
.send()
|
|
.await;
|
|
|
|
// Get last sync timestamp
|
|
let last_sync: Option<DateTime<Utc>> = sqlx::query_scalar(
|
|
"SELECT last_meili_sync FROM sync_metadata WHERE id = 1 AND last_meili_sync IS NOT NULL"
|
|
)
|
|
.fetch_optional(pool)
|
|
.await?;
|
|
|
|
// If no previous sync, do a full sync
|
|
let is_full_sync = last_sync.is_none();
|
|
|
|
// Get books to sync: all if full sync, only modified since last sync otherwise
|
|
let rows = if is_full_sync {
|
|
info!("[MEILI] Performing full sync");
|
|
sqlx::query(
|
|
"SELECT id, library_id, kind, title, author, series, volume, language, updated_at FROM books",
|
|
)
|
|
.fetch_all(pool)
|
|
.await?
|
|
} else {
|
|
let since = last_sync.unwrap();
|
|
info!("[MEILI] Performing incremental sync since {}", since);
|
|
|
|
// Also get deleted book IDs to remove from MeiliSearch
|
|
// For now, we'll do a diff approach: get all book IDs from DB and compare with Meili
|
|
sqlx::query(
|
|
"SELECT id, library_id, kind, title, author, series, volume, language, updated_at FROM books WHERE updated_at > $1",
|
|
)
|
|
.bind(since)
|
|
.fetch_all(pool)
|
|
.await?
|
|
};
|
|
|
|
if rows.is_empty() && !is_full_sync {
|
|
info!("[MEILI] No changes to sync");
|
|
// Still update the timestamp
|
|
sqlx::query(
|
|
"INSERT INTO sync_metadata (id, last_meili_sync) VALUES (1, NOW()) ON CONFLICT (id) DO UPDATE SET last_meili_sync = NOW()"
|
|
)
|
|
.execute(pool)
|
|
.await?;
|
|
return Ok(());
|
|
}
|
|
|
|
let docs: Vec<SearchDoc> = rows
|
|
.into_iter()
|
|
.map(|row| SearchDoc {
|
|
id: row.get::<Uuid, _>("id").to_string(),
|
|
library_id: row.get::<Uuid, _>("library_id").to_string(),
|
|
kind: row.get("kind"),
|
|
title: row.get("title"),
|
|
author: row.get("author"),
|
|
series: row.get("series"),
|
|
volume: row.get("volume"),
|
|
language: row.get("language"),
|
|
})
|
|
.collect();
|
|
|
|
let doc_count = docs.len();
|
|
|
|
// Send documents to MeiliSearch in batches of 1000
|
|
const MEILI_BATCH_SIZE: usize = 1000;
|
|
for (i, chunk) in docs.chunks(MEILI_BATCH_SIZE).enumerate() {
|
|
let batch_num = i + 1;
|
|
info!("[MEILI] Sending batch {}/{} ({} docs)", batch_num, doc_count.div_ceil(MEILI_BATCH_SIZE), chunk.len());
|
|
|
|
let response = client
|
|
.post(format!("{base}/indexes/books/documents"))
|
|
.header("Authorization", format!("Bearer {meili_master_key}"))
|
|
.json(&chunk)
|
|
.send()
|
|
.await
|
|
.context("failed to send docs to meili")?;
|
|
|
|
if !response.status().is_success() {
|
|
let status = response.status();
|
|
let body = response.text().await.unwrap_or_default();
|
|
return Err(anyhow::anyhow!("MeiliSearch error {}: {}", status, body));
|
|
}
|
|
}
|
|
|
|
// Handle deletions: get all book IDs from DB and remove from MeiliSearch any that don't exist
|
|
// This is expensive, so we only do it periodically (every 10 syncs) or on full syncs
|
|
if is_full_sync || rand::random::<u8>() < 26 { // ~10% chance
|
|
info!("[MEILI] Checking for documents to delete");
|
|
|
|
// Get all book IDs from database
|
|
let db_ids: Vec<String> = sqlx::query_scalar("SELECT id::text FROM books")
|
|
.fetch_all(pool)
|
|
.await?;
|
|
|
|
// Get all document IDs from MeiliSearch (this requires fetching all documents)
|
|
// For efficiency, we'll just delete by query for documents that might be stale
|
|
// A better approach would be to track deletions in a separate table
|
|
|
|
// For now, we'll do a simple approach: fetch all Meili docs and compare
|
|
// Note: This could be slow for large collections
|
|
let meili_response = client
|
|
.post(format!("{base}/indexes/books/documents/fetch"))
|
|
.header("Authorization", format!("Bearer {meili_master_key}"))
|
|
.json(&serde_json::json!({
|
|
"fields": ["id"],
|
|
"limit": 100000
|
|
}))
|
|
.send()
|
|
.await;
|
|
|
|
if let Ok(response) = meili_response {
|
|
if response.status().is_success() {
|
|
if let Ok(meili_docs) = response.json::<Vec<serde_json::Value>>().await {
|
|
let meili_ids: std::collections::HashSet<String> = meili_docs
|
|
.into_iter()
|
|
.filter_map(|doc| doc.get("id").and_then(|id| id.as_str()).map(|s| s.to_string()))
|
|
.collect();
|
|
|
|
let db_ids_set: std::collections::HashSet<String> = db_ids.into_iter().collect();
|
|
let to_delete: Vec<String> = meili_ids.difference(&db_ids_set).cloned().collect();
|
|
|
|
if !to_delete.is_empty() {
|
|
info!("[MEILI] Deleting {} stale documents", to_delete.len());
|
|
let _ = client
|
|
.post(format!("{base}/indexes/books/documents/delete-batch"))
|
|
.header("Authorization", format!("Bearer {meili_master_key}"))
|
|
.json(&to_delete)
|
|
.send()
|
|
.await;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Update last sync timestamp
|
|
sqlx::query(
|
|
"INSERT INTO sync_metadata (id, last_meili_sync) VALUES (1, NOW()) ON CONFLICT (id) DO UPDATE SET last_meili_sync = NOW()"
|
|
)
|
|
.execute(pool)
|
|
.await?;
|
|
|
|
info!("[MEILI] Sync completed: {} documents indexed", doc_count);
|
|
Ok(())
|
|
}
|