Migrate download queue from JSON to SQLite database
- Created QueueRepository adapter in src/server/services/queue_repository.py - Refactored DownloadService to use repository pattern instead of JSON - Updated application startup to initialize download service from database - Updated all test fixtures to use MockQueueRepository - All 1104 tests passing
This commit is contained in:
@@ -264,3 +264,35 @@ def get_sync_session() -> Session:
|
||||
)
|
||||
|
||||
return _sync_session_factory()
|
||||
|
||||
|
||||
def get_async_session_factory() -> AsyncSession:
|
||||
"""Get a new async database session (factory function).
|
||||
|
||||
Creates a new session instance for use in repository patterns.
|
||||
The caller is responsible for committing/rolling back and closing.
|
||||
|
||||
Returns:
|
||||
AsyncSession: New database session for async operations
|
||||
|
||||
Raises:
|
||||
RuntimeError: If database is not initialized
|
||||
|
||||
Example:
|
||||
session = get_async_session_factory()
|
||||
try:
|
||||
result = await session.execute(select(AnimeSeries))
|
||||
await session.commit()
|
||||
return result.scalars().all()
|
||||
except Exception:
|
||||
await session.rollback()
|
||||
raise
|
||||
finally:
|
||||
await session.close()
|
||||
"""
|
||||
if _session_factory is None:
|
||||
raise RuntimeError(
|
||||
"Database not initialized. Call init_db() first."
|
||||
)
|
||||
|
||||
return _session_factory()
|
||||
|
||||
@@ -112,6 +112,16 @@ async def lifespan(app: FastAPI):
|
||||
# Subscribe to progress events
|
||||
progress_service.subscribe("progress_updated", progress_event_handler)
|
||||
|
||||
# Initialize download service and restore queue from database
|
||||
try:
|
||||
from src.server.utils.dependencies import get_download_service
|
||||
download_service = get_download_service()
|
||||
await download_service.initialize()
|
||||
logger.info("Download service initialized and queue restored")
|
||||
except Exception as e:
|
||||
logger.warning("Failed to initialize download service: %s", e)
|
||||
# Continue startup - download service can be initialized later
|
||||
|
||||
logger.info("FastAPI application started successfully")
|
||||
logger.info("Server running on http://127.0.0.1:8000")
|
||||
logger.info(
|
||||
|
||||
@@ -2,18 +2,19 @@
|
||||
|
||||
This module provides a simplified queue management system for handling
|
||||
anime episode downloads with manual start/stop controls, progress tracking,
|
||||
persistence, and retry functionality.
|
||||
database persistence, and retry functionality.
|
||||
|
||||
The service uses SQLite database for persistent storage via QueueRepository
|
||||
while maintaining an in-memory cache for performance.
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
import asyncio
|
||||
import json
|
||||
import uuid
|
||||
from collections import deque
|
||||
from concurrent.futures import ThreadPoolExecutor
|
||||
from datetime import datetime, timezone
|
||||
from pathlib import Path
|
||||
from typing import Dict, List, Optional
|
||||
from typing import TYPE_CHECKING, Dict, List, Optional
|
||||
|
||||
import structlog
|
||||
|
||||
@@ -28,6 +29,9 @@ from src.server.models.download import (
|
||||
from src.server.services.anime_service import AnimeService, AnimeServiceError
|
||||
from src.server.services.progress_service import ProgressService, get_progress_service
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from src.server.services.queue_repository import QueueRepository
|
||||
|
||||
logger = structlog.get_logger(__name__)
|
||||
|
||||
|
||||
@@ -42,7 +46,7 @@ class DownloadService:
|
||||
- Manual download start/stop
|
||||
- FIFO queue processing
|
||||
- Real-time progress tracking
|
||||
- Queue persistence and recovery
|
||||
- Database persistence via QueueRepository
|
||||
- Automatic retry logic
|
||||
- WebSocket broadcast support
|
||||
"""
|
||||
@@ -50,24 +54,28 @@ class DownloadService:
|
||||
def __init__(
|
||||
self,
|
||||
anime_service: AnimeService,
|
||||
queue_repository: Optional["QueueRepository"] = None,
|
||||
max_retries: int = 3,
|
||||
persistence_path: str = "./data/download_queue.json",
|
||||
progress_service: Optional[ProgressService] = None,
|
||||
):
|
||||
"""Initialize the download service.
|
||||
|
||||
Args:
|
||||
anime_service: Service for anime operations
|
||||
queue_repository: Optional repository for database persistence.
|
||||
If not provided, will use default singleton.
|
||||
max_retries: Maximum retry attempts for failed downloads
|
||||
persistence_path: Path to persist queue state
|
||||
progress_service: Optional progress service for tracking
|
||||
"""
|
||||
self._anime_service = anime_service
|
||||
self._max_retries = max_retries
|
||||
self._persistence_path = Path(persistence_path)
|
||||
self._progress_service = progress_service or get_progress_service()
|
||||
|
||||
# Database repository for persistence
|
||||
self._queue_repository = queue_repository
|
||||
self._db_initialized = False
|
||||
|
||||
# Queue storage by status
|
||||
# In-memory cache for performance (synced with database)
|
||||
self._pending_queue: deque[DownloadItem] = deque()
|
||||
# Helper dict for O(1) lookup of pending items by ID
|
||||
self._pending_items_by_id: Dict[str, DownloadItem] = {}
|
||||
@@ -92,14 +100,158 @@ class DownloadService:
|
||||
# Track if queue progress has been initialized
|
||||
self._queue_progress_initialized: bool = False
|
||||
|
||||
# Load persisted queue
|
||||
self._load_queue()
|
||||
|
||||
logger.info(
|
||||
"DownloadService initialized",
|
||||
max_retries=max_retries,
|
||||
)
|
||||
|
||||
def _get_repository(self) -> "QueueRepository":
|
||||
"""Get the queue repository, initializing if needed.
|
||||
|
||||
Returns:
|
||||
QueueRepository instance
|
||||
"""
|
||||
if self._queue_repository is None:
|
||||
from src.server.services.queue_repository import get_queue_repository
|
||||
self._queue_repository = get_queue_repository()
|
||||
return self._queue_repository
|
||||
|
||||
async def initialize(self) -> None:
|
||||
"""Initialize the service by loading queue state from database.
|
||||
|
||||
Should be called after database is initialized during app startup.
|
||||
"""
|
||||
if self._db_initialized:
|
||||
return
|
||||
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
|
||||
# Load pending items from database
|
||||
pending_items = await repository.get_pending_items()
|
||||
for item in pending_items:
|
||||
# Reset status if was downloading when saved
|
||||
if item.status == DownloadStatus.DOWNLOADING:
|
||||
item.status = DownloadStatus.PENDING
|
||||
await repository.update_status(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
self._add_to_pending_queue(item)
|
||||
|
||||
# Load failed items from database
|
||||
failed_items = await repository.get_failed_items()
|
||||
for item in failed_items:
|
||||
if item.retry_count < self._max_retries:
|
||||
item.status = DownloadStatus.PENDING
|
||||
await repository.update_status(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
self._add_to_pending_queue(item)
|
||||
else:
|
||||
self._failed_items.append(item)
|
||||
|
||||
# Load completed items for history
|
||||
completed_items = await repository.get_completed_items(limit=100)
|
||||
for item in completed_items:
|
||||
self._completed_items.append(item)
|
||||
|
||||
self._db_initialized = True
|
||||
|
||||
logger.info(
|
||||
"Queue restored from database",
|
||||
pending_count=len(self._pending_queue),
|
||||
failed_count=len(self._failed_items),
|
||||
completed_count=len(self._completed_items),
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error("Failed to load queue from database", error=str(e))
|
||||
# Continue without persistence - queue will work in memory only
|
||||
self._db_initialized = True
|
||||
|
||||
async def _save_to_database(self, item: DownloadItem) -> DownloadItem:
|
||||
"""Save or update an item in the database.
|
||||
|
||||
Args:
|
||||
item: Download item to save
|
||||
|
||||
Returns:
|
||||
Saved item with database ID
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.save_item(item)
|
||||
except Exception as e:
|
||||
logger.error("Failed to save item to database", error=str(e))
|
||||
return item
|
||||
|
||||
async def _update_status_in_database(
|
||||
self,
|
||||
item_id: str,
|
||||
status: DownloadStatus,
|
||||
error: Optional[str] = None,
|
||||
) -> bool:
|
||||
"""Update item status in the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
status: New status
|
||||
error: Optional error message
|
||||
|
||||
Returns:
|
||||
True if update succeeded
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.update_status(item_id, status, error)
|
||||
except Exception as e:
|
||||
logger.error("Failed to update status in database", error=str(e))
|
||||
return False
|
||||
|
||||
async def _update_progress_in_database(
|
||||
self,
|
||||
item_id: str,
|
||||
progress: float,
|
||||
downloaded: int,
|
||||
total: Optional[int],
|
||||
speed: Optional[float],
|
||||
) -> bool:
|
||||
"""Update download progress in the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
progress: Progress percentage
|
||||
downloaded: Downloaded bytes
|
||||
total: Total bytes
|
||||
speed: Download speed in bytes/sec
|
||||
|
||||
Returns:
|
||||
True if update succeeded
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.update_progress(
|
||||
item_id, progress, downloaded, total, speed
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error("Failed to update progress in database", error=str(e))
|
||||
return False
|
||||
|
||||
async def _delete_from_database(self, item_id: str) -> bool:
|
||||
"""Delete an item from the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
|
||||
Returns:
|
||||
True if delete succeeded
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.delete_item(item_id)
|
||||
except Exception as e:
|
||||
logger.error("Failed to delete from database", error=str(e))
|
||||
return False
|
||||
|
||||
async def _init_queue_progress(self) -> None:
|
||||
"""Initialize the download queue progress tracking.
|
||||
|
||||
@@ -165,69 +317,6 @@ class DownloadService:
|
||||
"""Generate unique identifier for download items."""
|
||||
return str(uuid.uuid4())
|
||||
|
||||
def _load_queue(self) -> None:
|
||||
"""Load persisted queue from disk."""
|
||||
try:
|
||||
if self._persistence_path.exists():
|
||||
with open(self._persistence_path, "r", encoding="utf-8") as f:
|
||||
data = json.load(f)
|
||||
|
||||
# Restore pending items
|
||||
for item_dict in data.get("pending", []):
|
||||
item = DownloadItem(**item_dict)
|
||||
# Reset status if was downloading when saved
|
||||
if item.status == DownloadStatus.DOWNLOADING:
|
||||
item.status = DownloadStatus.PENDING
|
||||
self._add_to_pending_queue(item)
|
||||
|
||||
# Restore failed items that can be retried
|
||||
for item_dict in data.get("failed", []):
|
||||
item = DownloadItem(**item_dict)
|
||||
if item.retry_count < self._max_retries:
|
||||
item.status = DownloadStatus.PENDING
|
||||
self._add_to_pending_queue(item)
|
||||
else:
|
||||
self._failed_items.append(item)
|
||||
|
||||
logger.info(
|
||||
"Queue restored from disk",
|
||||
pending_count=len(self._pending_queue),
|
||||
failed_count=len(self._failed_items),
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error("Failed to load persisted queue", error=str(e))
|
||||
|
||||
def _save_queue(self) -> None:
|
||||
"""Persist current queue state to disk."""
|
||||
try:
|
||||
self._persistence_path.parent.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
active_items = (
|
||||
[self._active_download] if self._active_download else []
|
||||
)
|
||||
|
||||
data = {
|
||||
"pending": [
|
||||
item.model_dump(mode="json")
|
||||
for item in self._pending_queue
|
||||
],
|
||||
"active": [
|
||||
item.model_dump(mode="json") for item in active_items
|
||||
],
|
||||
"failed": [
|
||||
item.model_dump(mode="json")
|
||||
for item in self._failed_items
|
||||
],
|
||||
"timestamp": datetime.now(timezone.utc).isoformat(),
|
||||
}
|
||||
|
||||
with open(self._persistence_path, "w", encoding="utf-8") as f:
|
||||
json.dump(data, f, indent=2)
|
||||
|
||||
logger.debug("Queue persisted to disk")
|
||||
except Exception as e:
|
||||
logger.error("Failed to persist queue", error=str(e))
|
||||
|
||||
async def add_to_queue(
|
||||
self,
|
||||
serie_id: str,
|
||||
@@ -274,22 +363,23 @@ class DownloadService:
|
||||
added_at=datetime.now(timezone.utc),
|
||||
)
|
||||
|
||||
# Always append to end (FIFO order)
|
||||
self._add_to_pending_queue(item, front=False)
|
||||
# Save to database first to get persistent ID
|
||||
saved_item = await self._save_to_database(item)
|
||||
|
||||
created_ids.append(item.id)
|
||||
# Add to in-memory cache
|
||||
self._add_to_pending_queue(saved_item, front=False)
|
||||
|
||||
created_ids.append(saved_item.id)
|
||||
|
||||
logger.info(
|
||||
"Item added to queue",
|
||||
item_id=item.id,
|
||||
item_id=saved_item.id,
|
||||
serie_key=serie_id,
|
||||
serie_name=serie_name,
|
||||
season=episode.season,
|
||||
episode=episode.episode,
|
||||
)
|
||||
|
||||
self._save_queue()
|
||||
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@@ -333,6 +423,10 @@ class DownloadService:
|
||||
item.completed_at = datetime.now(timezone.utc)
|
||||
self._failed_items.append(item)
|
||||
self._active_download = None
|
||||
# Update status in database
|
||||
await self._update_status_in_database(
|
||||
item_id, DownloadStatus.CANCELLED
|
||||
)
|
||||
removed_ids.append(item_id)
|
||||
logger.info("Cancelled active download", item_id=item_id)
|
||||
continue
|
||||
@@ -342,13 +436,14 @@ class DownloadService:
|
||||
item = self._pending_items_by_id[item_id]
|
||||
self._pending_queue.remove(item)
|
||||
del self._pending_items_by_id[item_id]
|
||||
# Delete from database
|
||||
await self._delete_from_database(item_id)
|
||||
removed_ids.append(item_id)
|
||||
logger.info(
|
||||
"Removed from pending queue", item_id=item_id
|
||||
)
|
||||
|
||||
if removed_ids:
|
||||
self._save_queue()
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@@ -379,6 +474,10 @@ class DownloadService:
|
||||
|
||||
Raises:
|
||||
DownloadServiceError: If reordering fails
|
||||
|
||||
Note:
|
||||
Reordering is done in-memory only. Database priority is not
|
||||
updated since the in-memory queue defines the actual order.
|
||||
"""
|
||||
try:
|
||||
# Build new queue based on specified order
|
||||
@@ -399,9 +498,6 @@ class DownloadService:
|
||||
# Replace queue
|
||||
self._pending_queue = new_queue
|
||||
|
||||
# Save updated queue
|
||||
self._save_queue()
|
||||
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@@ -692,13 +788,15 @@ class DownloadService:
|
||||
Number of items cleared
|
||||
"""
|
||||
count = len(self._pending_queue)
|
||||
|
||||
# Delete all pending items from database
|
||||
for item_id in list(self._pending_items_by_id.keys()):
|
||||
await self._delete_from_database(item_id)
|
||||
|
||||
self._pending_queue.clear()
|
||||
self._pending_items_by_id.clear()
|
||||
logger.info("Cleared pending items", count=count)
|
||||
|
||||
# Save queue state
|
||||
self._save_queue()
|
||||
|
||||
# Notify via progress service
|
||||
if count > 0:
|
||||
queue_status = await self.get_queue_status()
|
||||
@@ -749,6 +847,11 @@ class DownloadService:
|
||||
self._add_to_pending_queue(item)
|
||||
retried_ids.append(item.id)
|
||||
|
||||
# Update status in database
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
|
||||
logger.info(
|
||||
"Retrying failed item",
|
||||
item_id=item.id,
|
||||
@@ -756,7 +859,6 @@ class DownloadService:
|
||||
)
|
||||
|
||||
if retried_ids:
|
||||
self._save_queue()
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@@ -790,10 +892,13 @@ class DownloadService:
|
||||
logger.info("Skipping download due to shutdown")
|
||||
return
|
||||
|
||||
# Update status
|
||||
# Update status in memory and database
|
||||
item.status = DownloadStatus.DOWNLOADING
|
||||
item.started_at = datetime.now(timezone.utc)
|
||||
self._active_download = item
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.DOWNLOADING
|
||||
)
|
||||
|
||||
logger.info(
|
||||
"Starting download",
|
||||
@@ -809,7 +914,8 @@ class DownloadService:
|
||||
# - download started/progress/completed/failed events
|
||||
# - All updates forwarded to ProgressService
|
||||
# - ProgressService broadcasts to WebSocket clients
|
||||
# Use serie_folder for filesystem operations and serie_id (key) for identification
|
||||
# Use serie_folder for filesystem operations
|
||||
# and serie_id (key) for identification
|
||||
if not item.serie_folder:
|
||||
raise DownloadServiceError(
|
||||
f"Missing serie_folder for download item {item.id}. "
|
||||
@@ -835,6 +941,11 @@ class DownloadService:
|
||||
|
||||
self._completed_items.append(item)
|
||||
|
||||
# Update database
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.COMPLETED
|
||||
)
|
||||
|
||||
logger.info(
|
||||
"Download completed successfully", item_id=item.id
|
||||
)
|
||||
@@ -849,9 +960,15 @@ class DownloadService:
|
||||
)
|
||||
item.status = DownloadStatus.CANCELLED
|
||||
item.completed_at = datetime.now(timezone.utc)
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.CANCELLED
|
||||
)
|
||||
# Return item to pending queue if not shutting down
|
||||
if not self._is_shutting_down:
|
||||
self._add_to_pending_queue(item, front=True)
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
raise # Re-raise to properly cancel the task
|
||||
|
||||
except Exception as e:
|
||||
@@ -861,6 +978,11 @@ class DownloadService:
|
||||
item.error = str(e)
|
||||
self._failed_items.append(item)
|
||||
|
||||
# Update database with error
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.FAILED, str(e)
|
||||
)
|
||||
|
||||
logger.error(
|
||||
"Download failed",
|
||||
item_id=item.id,
|
||||
@@ -874,8 +996,6 @@ class DownloadService:
|
||||
# Remove from active downloads
|
||||
if self._active_download and self._active_download.id == item.id:
|
||||
self._active_download = None
|
||||
|
||||
self._save_queue()
|
||||
|
||||
async def start(self) -> None:
|
||||
"""Initialize the download queue service (compatibility method).
|
||||
@@ -896,17 +1016,15 @@ class DownloadService:
|
||||
self._is_stopped = True
|
||||
|
||||
# Cancel active download task if running
|
||||
if self._active_download_task and not self._active_download_task.done():
|
||||
active_task = self._active_download_task
|
||||
if active_task and not active_task.done():
|
||||
logger.info("Cancelling active download task...")
|
||||
self._active_download_task.cancel()
|
||||
active_task.cancel()
|
||||
try:
|
||||
await self._active_download_task
|
||||
await active_task
|
||||
except asyncio.CancelledError:
|
||||
logger.info("Active download task cancelled")
|
||||
|
||||
# Save final state
|
||||
self._save_queue()
|
||||
|
||||
# Shutdown executor immediately, don't wait for tasks
|
||||
logger.info("Shutting down thread pool executor...")
|
||||
self._executor.shutdown(wait=False, cancel_futures=True)
|
||||
|
||||
753
src/server/services/queue_repository.py
Normal file
753
src/server/services/queue_repository.py
Normal file
@@ -0,0 +1,753 @@
|
||||
"""Queue repository adapter for database-backed download queue operations.
|
||||
|
||||
This module provides a repository adapter that wraps the DownloadQueueService
|
||||
and provides the interface needed by DownloadService for queue persistence.
|
||||
|
||||
The repository pattern abstracts the database operations from the business logic,
|
||||
allowing the DownloadService to work with domain models (DownloadItem) while
|
||||
the repository handles conversion to/from database models (DownloadQueueItem).
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
import logging
|
||||
from datetime import datetime, timezone
|
||||
from typing import Callable, List, Optional
|
||||
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
|
||||
from src.server.database.models import AnimeSeries
|
||||
from src.server.database.models import DownloadPriority as DBDownloadPriority
|
||||
from src.server.database.models import DownloadQueueItem as DBDownloadQueueItem
|
||||
from src.server.database.models import DownloadStatus as DBDownloadStatus
|
||||
from src.server.database.service import AnimeSeriesService, DownloadQueueService
|
||||
from src.server.models.download import (
|
||||
DownloadItem,
|
||||
DownloadPriority,
|
||||
DownloadProgress,
|
||||
DownloadStatus,
|
||||
EpisodeIdentifier,
|
||||
)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class QueueRepositoryError(Exception):
|
||||
"""Repository-level exception for queue operations."""
|
||||
|
||||
|
||||
class QueueRepository:
|
||||
"""Repository adapter for database-backed download queue operations.
|
||||
|
||||
Provides clean interface for queue operations while handling
|
||||
model conversion between Pydantic (DownloadItem) and SQLAlchemy
|
||||
(DownloadQueueItem) models.
|
||||
|
||||
Attributes:
|
||||
_db_session_factory: Factory function to create database sessions
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
db_session_factory: Callable[[], AsyncSession],
|
||||
) -> None:
|
||||
"""Initialize the queue repository.
|
||||
|
||||
Args:
|
||||
db_session_factory: Factory function that returns AsyncSession instances
|
||||
"""
|
||||
self._db_session_factory = db_session_factory
|
||||
logger.info("QueueRepository initialized")
|
||||
|
||||
# =========================================================================
|
||||
# Model Conversion Methods
|
||||
# =========================================================================
|
||||
|
||||
def _status_to_db(self, status: DownloadStatus) -> DBDownloadStatus:
|
||||
"""Convert Pydantic DownloadStatus to SQLAlchemy DownloadStatus.
|
||||
|
||||
Args:
|
||||
status: Pydantic status enum
|
||||
|
||||
Returns:
|
||||
SQLAlchemy status enum
|
||||
"""
|
||||
return DBDownloadStatus(status.value)
|
||||
|
||||
def _status_from_db(self, status: DBDownloadStatus) -> DownloadStatus:
|
||||
"""Convert SQLAlchemy DownloadStatus to Pydantic DownloadStatus.
|
||||
|
||||
Args:
|
||||
status: SQLAlchemy status enum
|
||||
|
||||
Returns:
|
||||
Pydantic status enum
|
||||
"""
|
||||
return DownloadStatus(status.value)
|
||||
|
||||
def _priority_to_db(self, priority: DownloadPriority) -> DBDownloadPriority:
|
||||
"""Convert Pydantic DownloadPriority to SQLAlchemy DownloadPriority.
|
||||
|
||||
Args:
|
||||
priority: Pydantic priority enum
|
||||
|
||||
Returns:
|
||||
SQLAlchemy priority enum
|
||||
"""
|
||||
# Handle case differences (Pydantic uses uppercase, DB uses lowercase)
|
||||
return DBDownloadPriority(priority.value.lower())
|
||||
|
||||
def _priority_from_db(self, priority: DBDownloadPriority) -> DownloadPriority:
|
||||
"""Convert SQLAlchemy DownloadPriority to Pydantic DownloadPriority.
|
||||
|
||||
Args:
|
||||
priority: SQLAlchemy priority enum
|
||||
|
||||
Returns:
|
||||
Pydantic priority enum
|
||||
"""
|
||||
# Handle case differences (DB uses lowercase, Pydantic uses uppercase)
|
||||
return DownloadPriority(priority.value.upper())
|
||||
|
||||
def _to_db_model(
|
||||
self,
|
||||
item: DownloadItem,
|
||||
series_id: int,
|
||||
) -> DBDownloadQueueItem:
|
||||
"""Convert DownloadItem to database model.
|
||||
|
||||
Args:
|
||||
item: Pydantic download item
|
||||
series_id: Database series ID (foreign key)
|
||||
|
||||
Returns:
|
||||
SQLAlchemy download queue item model
|
||||
"""
|
||||
return DBDownloadQueueItem(
|
||||
series_id=series_id,
|
||||
season=item.episode.season,
|
||||
episode_number=item.episode.episode,
|
||||
status=self._status_to_db(item.status),
|
||||
priority=self._priority_to_db(item.priority),
|
||||
progress_percent=item.progress.percent if item.progress else 0.0,
|
||||
downloaded_bytes=int(
|
||||
item.progress.downloaded_mb * 1024 * 1024
|
||||
) if item.progress else 0,
|
||||
total_bytes=int(
|
||||
item.progress.total_mb * 1024 * 1024
|
||||
) if item.progress and item.progress.total_mb else None,
|
||||
download_speed=(
|
||||
item.progress.speed_mbps * 1024 * 1024
|
||||
) if item.progress and item.progress.speed_mbps else None,
|
||||
error_message=item.error,
|
||||
retry_count=item.retry_count,
|
||||
download_url=str(item.source_url) if item.source_url else None,
|
||||
started_at=item.started_at,
|
||||
completed_at=item.completed_at,
|
||||
)
|
||||
|
||||
def _from_db_model(
|
||||
self,
|
||||
db_item: DBDownloadQueueItem,
|
||||
item_id: Optional[str] = None,
|
||||
) -> DownloadItem:
|
||||
"""Convert database model to DownloadItem.
|
||||
|
||||
Args:
|
||||
db_item: SQLAlchemy download queue item
|
||||
item_id: Optional override for item ID (uses db ID if not provided)
|
||||
|
||||
Returns:
|
||||
Pydantic download item
|
||||
"""
|
||||
# Build progress object if there's progress data
|
||||
progress = None
|
||||
if db_item.progress_percent > 0 or db_item.downloaded_bytes > 0:
|
||||
progress = DownloadProgress(
|
||||
percent=db_item.progress_percent,
|
||||
downloaded_mb=db_item.downloaded_bytes / (1024 * 1024),
|
||||
total_mb=(
|
||||
db_item.total_bytes / (1024 * 1024)
|
||||
if db_item.total_bytes else None
|
||||
),
|
||||
speed_mbps=(
|
||||
db_item.download_speed / (1024 * 1024)
|
||||
if db_item.download_speed else None
|
||||
),
|
||||
)
|
||||
|
||||
return DownloadItem(
|
||||
id=item_id or str(db_item.id),
|
||||
serie_id=db_item.series.key if db_item.series else "",
|
||||
serie_folder=db_item.series.folder if db_item.series else "",
|
||||
serie_name=db_item.series.name if db_item.series else "",
|
||||
episode=EpisodeIdentifier(
|
||||
season=db_item.season,
|
||||
episode=db_item.episode_number,
|
||||
),
|
||||
status=self._status_from_db(db_item.status),
|
||||
priority=self._priority_from_db(db_item.priority),
|
||||
added_at=db_item.created_at or datetime.now(timezone.utc),
|
||||
started_at=db_item.started_at,
|
||||
completed_at=db_item.completed_at,
|
||||
progress=progress,
|
||||
error=db_item.error_message,
|
||||
retry_count=db_item.retry_count,
|
||||
source_url=db_item.download_url,
|
||||
)
|
||||
|
||||
# =========================================================================
|
||||
# CRUD Operations
|
||||
# =========================================================================
|
||||
|
||||
async def save_item(
|
||||
self,
|
||||
item: DownloadItem,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> DownloadItem:
|
||||
"""Save a download item to the database.
|
||||
|
||||
Creates a new record if the item doesn't exist in the database.
|
||||
|
||||
Args:
|
||||
item: Download item to save
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Saved download item with database ID
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If save operation fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
# Find series by key
|
||||
series = await AnimeSeriesService.get_by_key(session, item.serie_id)
|
||||
|
||||
if not series:
|
||||
# Create series if it doesn't exist
|
||||
series = await AnimeSeriesService.create(
|
||||
db=session,
|
||||
key=item.serie_id,
|
||||
name=item.serie_name,
|
||||
site="", # Will be updated later if needed
|
||||
folder=item.serie_folder,
|
||||
)
|
||||
logger.info(
|
||||
"Created new series for queue item",
|
||||
key=item.serie_id,
|
||||
name=item.serie_name,
|
||||
)
|
||||
|
||||
# Create queue item
|
||||
db_item = await DownloadQueueService.create(
|
||||
db=session,
|
||||
series_id=series.id,
|
||||
season=item.episode.season,
|
||||
episode_number=item.episode.episode,
|
||||
priority=self._priority_to_db(item.priority),
|
||||
download_url=str(item.source_url) if item.source_url else None,
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
# Update the item ID with the database ID
|
||||
item.id = str(db_item.id)
|
||||
|
||||
logger.debug(
|
||||
"Saved queue item to database",
|
||||
item_id=item.id,
|
||||
serie_key=item.serie_id,
|
||||
)
|
||||
|
||||
return item
|
||||
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to save queue item", error=str(e))
|
||||
raise QueueRepositoryError(f"Failed to save item: {str(e)}") from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_item(
|
||||
self,
|
||||
item_id: str,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> Optional[DownloadItem]:
|
||||
"""Get a download item by ID.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID (database ID as string)
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Download item or None if not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_item = await DownloadQueueService.get_by_id(
|
||||
session, int(item_id)
|
||||
)
|
||||
|
||||
if not db_item:
|
||||
return None
|
||||
|
||||
return self._from_db_model(db_item, item_id)
|
||||
|
||||
except ValueError:
|
||||
# Invalid ID format
|
||||
return None
|
||||
except Exception as e:
|
||||
logger.error("Failed to get queue item", error=str(e))
|
||||
raise QueueRepositoryError(f"Failed to get item: {str(e)}") from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_pending_items(
|
||||
self,
|
||||
limit: Optional[int] = None,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get pending download items ordered by priority.
|
||||
|
||||
Args:
|
||||
limit: Optional maximum number of items to return
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of pending download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_pending(session, limit)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get pending items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get pending items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_active_item(
|
||||
self,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> Optional[DownloadItem]:
|
||||
"""Get the currently active (downloading) item.
|
||||
|
||||
Args:
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Active download item or None if none active
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_active(session)
|
||||
|
||||
if not db_items:
|
||||
return None
|
||||
|
||||
# Return first active item (should only be one)
|
||||
return self._from_db_model(db_items[0])
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get active item", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get active item: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_completed_items(
|
||||
self,
|
||||
limit: int = 100,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get completed download items.
|
||||
|
||||
Args:
|
||||
limit: Maximum number of items to return
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of completed download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_by_status(
|
||||
session, DBDownloadStatus.COMPLETED, limit
|
||||
)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get completed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get completed items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_failed_items(
|
||||
self,
|
||||
limit: int = 50,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get failed download items.
|
||||
|
||||
Args:
|
||||
limit: Maximum number of items to return
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of failed download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_by_status(
|
||||
session, DBDownloadStatus.FAILED, limit
|
||||
)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get failed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get failed items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def update_status(
|
||||
self,
|
||||
item_id: str,
|
||||
status: DownloadStatus,
|
||||
error: Optional[str] = None,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> bool:
|
||||
"""Update the status of a download item.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
status: New download status
|
||||
error: Optional error message for failed status
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
True if update succeeded, False if item not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If update fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
result = await DownloadQueueService.update_status(
|
||||
session,
|
||||
int(item_id),
|
||||
self._status_to_db(status),
|
||||
error,
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
success = result is not None
|
||||
|
||||
if success:
|
||||
logger.debug(
|
||||
"Updated queue item status",
|
||||
item_id=item_id,
|
||||
status=status.value,
|
||||
)
|
||||
|
||||
return success
|
||||
|
||||
except ValueError:
|
||||
return False
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to update status", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to update status: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def update_progress(
|
||||
self,
|
||||
item_id: str,
|
||||
progress: float,
|
||||
downloaded: int,
|
||||
total: Optional[int],
|
||||
speed: Optional[float],
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> bool:
|
||||
"""Update download progress for an item.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
progress: Progress percentage (0-100)
|
||||
downloaded: Downloaded bytes
|
||||
total: Total bytes (optional)
|
||||
speed: Download speed in bytes/second (optional)
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
True if update succeeded, False if item not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If update fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
result = await DownloadQueueService.update_progress(
|
||||
session,
|
||||
int(item_id),
|
||||
progress,
|
||||
downloaded,
|
||||
total,
|
||||
speed,
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
return result is not None
|
||||
|
||||
except ValueError:
|
||||
return False
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to update progress", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to update progress: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def delete_item(
|
||||
self,
|
||||
item_id: str,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> bool:
|
||||
"""Delete a download item from the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
True if item was deleted, False if not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If delete fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
result = await DownloadQueueService.delete(session, int(item_id))
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
if result:
|
||||
logger.debug("Deleted queue item", item_id=item_id)
|
||||
|
||||
return result
|
||||
|
||||
except ValueError:
|
||||
return False
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to delete item", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to delete item: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def clear_completed(
|
||||
self,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> int:
|
||||
"""Clear all completed download items.
|
||||
|
||||
Args:
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Number of items cleared
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If operation fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
count = await DownloadQueueService.clear_completed(session)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
logger.info("Cleared completed items from queue", count=count)
|
||||
return count
|
||||
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to clear completed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to clear completed: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_all_items(
|
||||
self,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get all download items regardless of status.
|
||||
|
||||
Args:
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of all download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_all(
|
||||
session, with_series=True
|
||||
)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get all items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get all items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def retry_failed_items(
|
||||
self,
|
||||
max_retries: int = 3,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Retry failed downloads that haven't exceeded max retries.
|
||||
|
||||
Args:
|
||||
max_retries: Maximum number of retry attempts
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of items marked for retry
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If operation fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.retry_failed(
|
||||
session, max_retries
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to retry failed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to retry failed items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
|
||||
# Singleton instance
|
||||
_queue_repository_instance: Optional[QueueRepository] = None
|
||||
|
||||
|
||||
def get_queue_repository(
|
||||
db_session_factory: Optional[Callable[[], AsyncSession]] = None,
|
||||
) -> QueueRepository:
|
||||
"""Get or create the QueueRepository singleton.
|
||||
|
||||
Args:
|
||||
db_session_factory: Optional factory function for database sessions.
|
||||
If not provided, uses default from connection module.
|
||||
|
||||
Returns:
|
||||
QueueRepository singleton instance
|
||||
"""
|
||||
global _queue_repository_instance
|
||||
|
||||
if _queue_repository_instance is None:
|
||||
if db_session_factory is None:
|
||||
# Use default session factory
|
||||
from src.server.database.connection import get_async_session_factory
|
||||
db_session_factory = get_async_session_factory
|
||||
|
||||
_queue_repository_instance = QueueRepository(db_session_factory)
|
||||
|
||||
return _queue_repository_instance
|
||||
Reference in New Issue
Block a user