feat: section disponibles au téléchargement + fix nommage import
All checks were successful
Deploy with Docker Compose / deploy (push) Successful in 43s
All checks were successful
Deploy with Docker Compose / deploy (push) Successful in 43s
- Endpoint GET /download-detection/latest-found : résultats "found" du dernier job de détection par bibliothèque - Section dans la page Téléchargements avec les releases disponibles groupées par bibliothèque, bouton qBittorrent intégré - Fix nommage import : exclut les volumes importés de la recherche de référence (évite le cercle vicieux vol 8 → ref vol 8 → même nom) - Fix extraction volumes : gère "Tome.007" (point après préfixe) en plus de "Tome 007" dans extract_volumes_from_title - Fallback disque pour la référence de nommage quand la DB ne matche pas - Logging détaillé du processus d'import pour debug Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
This commit is contained in:
@@ -343,6 +343,97 @@ pub async fn get_detection_results(
|
||||
Ok(Json(results))
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// GET /download-detection/latest-found
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
#[derive(Serialize, ToSchema)]
|
||||
pub struct LatestFoundPerLibraryDto {
|
||||
#[schema(value_type = String)]
|
||||
pub library_id: Uuid,
|
||||
pub library_name: String,
|
||||
#[schema(value_type = String)]
|
||||
pub job_id: Uuid,
|
||||
pub job_date: String,
|
||||
pub results: Vec<DownloadDetectionResultDto>,
|
||||
}
|
||||
|
||||
/// Returns "found" results from the latest detection job per library.
|
||||
#[utoipa::path(
|
||||
get,
|
||||
path = "/download-detection/latest-found",
|
||||
tag = "download_detection",
|
||||
responses(
|
||||
(status = 200, body = Vec<LatestFoundPerLibraryDto>),
|
||||
),
|
||||
security(("Bearer" = []))
|
||||
)]
|
||||
pub async fn get_latest_found(
|
||||
State(state): State<AppState>,
|
||||
) -> Result<Json<Vec<LatestFoundPerLibraryDto>>, ApiError> {
|
||||
// Get latest completed detection job per library
|
||||
let jobs = sqlx::query(
|
||||
"SELECT DISTINCT ON (j.library_id) j.id, j.library_id, j.created_at, l.name as library_name \
|
||||
FROM index_jobs j \
|
||||
JOIN libraries l ON l.id = j.library_id \
|
||||
WHERE j.type = 'download_detection' AND j.status = 'success' \
|
||||
ORDER BY j.library_id, j.created_at DESC",
|
||||
)
|
||||
.fetch_all(&state.pool)
|
||||
.await?;
|
||||
|
||||
let mut output = Vec::new();
|
||||
|
||||
for job in &jobs {
|
||||
let job_id: Uuid = job.get("id");
|
||||
let library_id: Uuid = job.get("library_id");
|
||||
let library_name: String = job.get("library_name");
|
||||
let created_at: chrono::DateTime<chrono::Utc> = job.get("created_at");
|
||||
|
||||
let rows = sqlx::query(
|
||||
"SELECT id, series_name, status, missing_count, available_releases, error_message \
|
||||
FROM download_detection_results \
|
||||
WHERE job_id = $1 AND status = 'found' \
|
||||
ORDER BY series_name",
|
||||
)
|
||||
.bind(job_id)
|
||||
.fetch_all(&state.pool)
|
||||
.await?;
|
||||
|
||||
if rows.is_empty() {
|
||||
continue;
|
||||
}
|
||||
|
||||
let results = rows
|
||||
.iter()
|
||||
.map(|row| {
|
||||
let releases_json: Option<serde_json::Value> = row.get("available_releases");
|
||||
let available_releases = releases_json.and_then(|v| {
|
||||
serde_json::from_value::<Vec<AvailableReleaseDto>>(v).ok()
|
||||
});
|
||||
DownloadDetectionResultDto {
|
||||
id: row.get("id"),
|
||||
series_name: row.get("series_name"),
|
||||
status: row.get("status"),
|
||||
missing_count: row.get("missing_count"),
|
||||
available_releases,
|
||||
error_message: row.get("error_message"),
|
||||
}
|
||||
})
|
||||
.collect();
|
||||
|
||||
output.push(LatestFoundPerLibraryDto {
|
||||
library_id,
|
||||
library_name,
|
||||
job_id,
|
||||
job_date: created_at.to_rfc3339(),
|
||||
results,
|
||||
});
|
||||
}
|
||||
|
||||
Ok(Json(output))
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Background processing
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
Reference in New Issue
Block a user