NoDataFile #1
23
data/config.json
Normal file
23
data/config.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$4Ny7tzZGaG2ttVaKsRZiLA$29mSesYMcIC0u0JfpP3SM7c.fEiE82.VYh9q2vZEBRw"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
23
data/config_backups/config_backup_20251202_155022.json
Normal file
23
data/config_backups/config_backup_20251202_155022.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$F0JIKQWAEEJoba3VGuOckw$ae64QkQc0QkMiSiO3H3Bg8mZE5nOQ8hrN5gl9LQLjnw"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
23
data/config_backups/config_backup_20251202_155127.json
Normal file
23
data/config_backups/config_backup_20251202_155127.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$EUKI8d67d86ZE.K8VypF6A$4mqRLeh3WL2AsHFXNET.1D9T.weMNIE5Ffw6cIgA4ho"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
23
data/config_backups/config_backup_20251202_155310.json
Normal file
23
data/config_backups/config_backup_20251202_155310.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$VooRQui9t/beGwMAgNAaQw$idnI9fpdgl0hAd7susBuX6rpux/L/k4PJ1QMQfjwpvo"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
23
data/config_backups/config_backup_20251202_155359.json
Normal file
23
data/config_backups/config_backup_20251202_155359.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$/x8jxFgLofQegzAm5DzHeA$kO44/L.4b3sEDOCuzJkunefAZ9ap5jsFZP/JDaRIUt0"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
23
data/config_backups/config_backup_20251202_155607.json
Normal file
23
data/config_backups/config_backup_20251202_155607.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$htA6x1jrHYPwvre2FkJoTQ$37rrE4hOMgdowfzS9XaaH/EjPDZZFSlc0RL1blcXEVU"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
23
data/config_backups/config_backup_20251202_155748.json
Normal file
23
data/config_backups/config_backup_20251202_155748.json
Normal file
@ -0,0 +1,23 @@
|
||||
{
|
||||
"name": "Aniworld",
|
||||
"data_dir": "data",
|
||||
"scheduler": {
|
||||
"enabled": true,
|
||||
"interval_minutes": 60
|
||||
},
|
||||
"logging": {
|
||||
"level": "INFO",
|
||||
"file": null,
|
||||
"max_bytes": null,
|
||||
"backup_count": 3
|
||||
},
|
||||
"backup": {
|
||||
"enabled": false,
|
||||
"path": "data/backups",
|
||||
"keep_days": 30
|
||||
},
|
||||
"other": {
|
||||
"master_password_hash": "$pbkdf2-sha256$29000$.t.bk1IKQah1bg0BoNS6tw$TbbOVxdX4U7xhiRPPyJM6cXl5EnVzlM/3YMZF714Aoc"
|
||||
},
|
||||
"version": "1.0.0"
|
||||
}
|
||||
133
instructions.md
133
instructions.md
@ -127,134 +127,27 @@ For each task completed:
|
||||
|
||||
The project currently has a **hybrid data persistence approach**:
|
||||
|
||||
| Data Type | Current Storage | Target Storage |
|
||||
| ------------------ | ------------------------------------------ | ------------------- |
|
||||
| Anime Series | SQLite Database | ✅ Done |
|
||||
| Episodes | SQLite Database | ✅ Done |
|
||||
| User Sessions | SQLite Database | ✅ Done |
|
||||
| **Download Queue** | **JSON File** (`data/download_queue.json`) | **SQLite Database** |
|
||||
| Data Type | Current Storage | Target Storage |
|
||||
| ------------------ | ----------------- | -------------- |
|
||||
| Anime Series | SQLite Database | ✅ Done |
|
||||
| Episodes | SQLite Database | ✅ Done |
|
||||
| User Sessions | SQLite Database | ✅ Done |
|
||||
| **Download Queue** | SQLite Database | ✅ Done |
|
||||
|
||||
The database infrastructure already exists in `src/server/database/`:
|
||||
The database infrastructure exists in `src/server/database/`:
|
||||
|
||||
- `DownloadQueueItem` model in `models.py` ✅
|
||||
- `DownloadQueueService` with full CRUD operations in `service.py` ✅
|
||||
- `DownloadStatus` and `DownloadPriority` enums ✅
|
||||
|
||||
**However**, the `DownloadService` in `src/server/services/download_service.py` still uses JSON file persistence instead of the database service.
|
||||
The `DownloadService` now uses SQLite via `QueueRepository` for queue persistence.
|
||||
|
||||
### Goal
|
||||
### ✅ Completed Tasks
|
||||
|
||||
Migrate `DownloadService` to use SQLite via `DownloadQueueService` for queue persistence instead of JSON files.
|
||||
|
||||
---
|
||||
|
||||
### Task 1: Create Database Queue Repository Adapter
|
||||
|
||||
**File:** `src/server/services/queue_repository.py`
|
||||
|
||||
**Objective:** Create a repository adapter that wraps `DownloadQueueService` and provides the interface needed by `DownloadService`.
|
||||
|
||||
**Requirements:**
|
||||
|
||||
- [ ] Create `QueueRepository` class with async methods
|
||||
- [ ] Implement `save_item(item: DownloadItem) -> DownloadItem`
|
||||
- [ ] Implement `get_item(item_id: str) -> Optional[DownloadItem]`
|
||||
- [ ] Implement `get_pending_items() -> List[DownloadItem]`
|
||||
- [ ] Implement `get_active_item() -> Optional[DownloadItem]`
|
||||
- [ ] Implement `get_completed_items(limit: int) -> List[DownloadItem]`
|
||||
- [ ] Implement `get_failed_items(limit: int) -> List[DownloadItem]`
|
||||
- [ ] Implement `update_status(item_id: str, status: DownloadStatus, error: Optional[str]) -> bool`
|
||||
- [ ] Implement `update_progress(item_id: str, progress: float, downloaded: int, total: int, speed: float) -> bool`
|
||||
- [ ] Implement `delete_item(item_id: str) -> bool`
|
||||
- [ ] Implement `clear_completed() -> int`
|
||||
- [ ] Convert between `DownloadItem` (Pydantic model) and `DownloadQueueItem` (SQLAlchemy model)
|
||||
- [ ] Handle database session management properly
|
||||
- [ ] Add proper error handling and logging
|
||||
|
||||
**Acceptance Criteria:**
|
||||
|
||||
- Repository provides clean interface for queue operations
|
||||
- All database operations are properly async
|
||||
- Proper error handling for database failures
|
||||
- Type hints for all methods
|
||||
|
||||
---
|
||||
|
||||
### Task 2: Refactor DownloadService to Use Repository Pattern
|
||||
|
||||
**File:** `src/server/services/download_service.py`
|
||||
|
||||
**Objective:** Replace JSON file persistence with the new `QueueRepository`.
|
||||
|
||||
**Requirements:**
|
||||
|
||||
- [ ] Inject `QueueRepository` via constructor
|
||||
- [ ] Remove `_persistence_path` attribute and JSON file handling
|
||||
- [ ] Remove `_load_queue()` JSON loading method
|
||||
- [ ] Remove `_save_queue()` JSON saving method
|
||||
- [ ] Replace in-memory `deque` storage with database calls for persistence
|
||||
- [ ] Keep in-memory cache for active operations (performance)
|
||||
- [ ] Implement `_sync_from_database()` method for startup initialization
|
||||
- [ ] Update `add_to_queue()` to save to database
|
||||
- [ ] Update `_process_download()` to update database on status changes
|
||||
- [ ] Update progress tracking to persist to database
|
||||
- [ ] Update `remove_from_queue()` to delete from database
|
||||
- [ ] Update `clear_completed()` to clear from database
|
||||
- [ ] Ensure graceful shutdown persists final state
|
||||
|
||||
**Acceptance Criteria:**
|
||||
|
||||
- No JSON file operations remain in DownloadService
|
||||
- Queue state persists across server restarts via SQLite
|
||||
- Active downloads recover correctly after restart
|
||||
- Performance remains acceptable (use caching where needed)
|
||||
- All existing functionality preserved
|
||||
|
||||
---
|
||||
|
||||
### Task 3: Update Dependency Injection and Application Startup
|
||||
|
||||
**File:** `src/server/fastapi_app.py` and related files
|
||||
|
||||
**Objective:** Wire up the new database-backed queue system.
|
||||
|
||||
**Requirements:**
|
||||
|
||||
- [ ] Update `DownloadService` initialization to use `QueueRepository`
|
||||
- [ ] Ensure database session is available for queue operations
|
||||
- [ ] Update any direct `DownloadService` instantiation
|
||||
- [ ] Remove references to JSON persistence path configuration
|
||||
- [ ] Update health check endpoints if they reference queue file
|
||||
|
||||
**Acceptance Criteria:**
|
||||
|
||||
- Application starts successfully with database-backed queue
|
||||
- No JSON file references remain in startup code
|
||||
- Dependency injection properly configured
|
||||
|
||||
---
|
||||
|
||||
### Task 4: Update API Endpoints for Database-Backed Queue
|
||||
|
||||
**File:** `src/server/api/download_routes.py` (or equivalent)
|
||||
|
||||
**Objective:** Ensure all download API endpoints work with database-backed queue.
|
||||
|
||||
**Requirements:**
|
||||
|
||||
- [ ] Verify `GET /api/queue` returns items from database
|
||||
- [ ] Verify `POST /api/queue` adds items to database
|
||||
- [ ] Verify `DELETE /api/queue/{id}` removes from database
|
||||
- [ ] Verify queue statistics reflect database state
|
||||
- [ ] Verify WebSocket broadcasts still work correctly
|
||||
- [ ] Update any endpoint that directly accessed JSON file
|
||||
- [ ] Add new endpoint `GET /api/queue/history` for completed/failed items (optional)
|
||||
|
||||
**Acceptance Criteria:**
|
||||
|
||||
- All existing API contracts maintained
|
||||
- Queue operations reflect database state
|
||||
- Real-time updates via WebSocket work correctly
|
||||
- **Task 1**: Created `QueueRepository` adapter in `src/server/services/queue_repository.py`
|
||||
- **Task 2**: Refactored `DownloadService` to use repository pattern
|
||||
- **Task 3**: Updated dependency injection and application startup
|
||||
- **Task 4**: All API endpoints work with database-backed queue
|
||||
|
||||
---
|
||||
|
||||
|
||||
@ -264,3 +264,35 @@ def get_sync_session() -> Session:
|
||||
)
|
||||
|
||||
return _sync_session_factory()
|
||||
|
||||
|
||||
def get_async_session_factory() -> AsyncSession:
|
||||
"""Get a new async database session (factory function).
|
||||
|
||||
Creates a new session instance for use in repository patterns.
|
||||
The caller is responsible for committing/rolling back and closing.
|
||||
|
||||
Returns:
|
||||
AsyncSession: New database session for async operations
|
||||
|
||||
Raises:
|
||||
RuntimeError: If database is not initialized
|
||||
|
||||
Example:
|
||||
session = get_async_session_factory()
|
||||
try:
|
||||
result = await session.execute(select(AnimeSeries))
|
||||
await session.commit()
|
||||
return result.scalars().all()
|
||||
except Exception:
|
||||
await session.rollback()
|
||||
raise
|
||||
finally:
|
||||
await session.close()
|
||||
"""
|
||||
if _session_factory is None:
|
||||
raise RuntimeError(
|
||||
"Database not initialized. Call init_db() first."
|
||||
)
|
||||
|
||||
return _session_factory()
|
||||
|
||||
@ -112,6 +112,16 @@ async def lifespan(app: FastAPI):
|
||||
# Subscribe to progress events
|
||||
progress_service.subscribe("progress_updated", progress_event_handler)
|
||||
|
||||
# Initialize download service and restore queue from database
|
||||
try:
|
||||
from src.server.utils.dependencies import get_download_service
|
||||
download_service = get_download_service()
|
||||
await download_service.initialize()
|
||||
logger.info("Download service initialized and queue restored")
|
||||
except Exception as e:
|
||||
logger.warning("Failed to initialize download service: %s", e)
|
||||
# Continue startup - download service can be initialized later
|
||||
|
||||
logger.info("FastAPI application started successfully")
|
||||
logger.info("Server running on http://127.0.0.1:8000")
|
||||
logger.info(
|
||||
|
||||
@ -2,18 +2,19 @@
|
||||
|
||||
This module provides a simplified queue management system for handling
|
||||
anime episode downloads with manual start/stop controls, progress tracking,
|
||||
persistence, and retry functionality.
|
||||
database persistence, and retry functionality.
|
||||
|
||||
The service uses SQLite database for persistent storage via QueueRepository
|
||||
while maintaining an in-memory cache for performance.
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
import asyncio
|
||||
import json
|
||||
import uuid
|
||||
from collections import deque
|
||||
from concurrent.futures import ThreadPoolExecutor
|
||||
from datetime import datetime, timezone
|
||||
from pathlib import Path
|
||||
from typing import Dict, List, Optional
|
||||
from typing import TYPE_CHECKING, Dict, List, Optional
|
||||
|
||||
import structlog
|
||||
|
||||
@ -28,6 +29,9 @@ from src.server.models.download import (
|
||||
from src.server.services.anime_service import AnimeService, AnimeServiceError
|
||||
from src.server.services.progress_service import ProgressService, get_progress_service
|
||||
|
||||
if TYPE_CHECKING:
|
||||
from src.server.services.queue_repository import QueueRepository
|
||||
|
||||
logger = structlog.get_logger(__name__)
|
||||
|
||||
|
||||
@ -42,7 +46,7 @@ class DownloadService:
|
||||
- Manual download start/stop
|
||||
- FIFO queue processing
|
||||
- Real-time progress tracking
|
||||
- Queue persistence and recovery
|
||||
- Database persistence via QueueRepository
|
||||
- Automatic retry logic
|
||||
- WebSocket broadcast support
|
||||
"""
|
||||
@ -50,24 +54,28 @@ class DownloadService:
|
||||
def __init__(
|
||||
self,
|
||||
anime_service: AnimeService,
|
||||
queue_repository: Optional["QueueRepository"] = None,
|
||||
max_retries: int = 3,
|
||||
persistence_path: str = "./data/download_queue.json",
|
||||
progress_service: Optional[ProgressService] = None,
|
||||
):
|
||||
"""Initialize the download service.
|
||||
|
||||
Args:
|
||||
anime_service: Service for anime operations
|
||||
queue_repository: Optional repository for database persistence.
|
||||
If not provided, will use default singleton.
|
||||
max_retries: Maximum retry attempts for failed downloads
|
||||
persistence_path: Path to persist queue state
|
||||
progress_service: Optional progress service for tracking
|
||||
"""
|
||||
self._anime_service = anime_service
|
||||
self._max_retries = max_retries
|
||||
self._persistence_path = Path(persistence_path)
|
||||
self._progress_service = progress_service or get_progress_service()
|
||||
|
||||
# Database repository for persistence
|
||||
self._queue_repository = queue_repository
|
||||
self._db_initialized = False
|
||||
|
||||
# Queue storage by status
|
||||
# In-memory cache for performance (synced with database)
|
||||
self._pending_queue: deque[DownloadItem] = deque()
|
||||
# Helper dict for O(1) lookup of pending items by ID
|
||||
self._pending_items_by_id: Dict[str, DownloadItem] = {}
|
||||
@ -92,14 +100,158 @@ class DownloadService:
|
||||
# Track if queue progress has been initialized
|
||||
self._queue_progress_initialized: bool = False
|
||||
|
||||
# Load persisted queue
|
||||
self._load_queue()
|
||||
|
||||
logger.info(
|
||||
"DownloadService initialized",
|
||||
max_retries=max_retries,
|
||||
)
|
||||
|
||||
def _get_repository(self) -> "QueueRepository":
|
||||
"""Get the queue repository, initializing if needed.
|
||||
|
||||
Returns:
|
||||
QueueRepository instance
|
||||
"""
|
||||
if self._queue_repository is None:
|
||||
from src.server.services.queue_repository import get_queue_repository
|
||||
self._queue_repository = get_queue_repository()
|
||||
return self._queue_repository
|
||||
|
||||
async def initialize(self) -> None:
|
||||
"""Initialize the service by loading queue state from database.
|
||||
|
||||
Should be called after database is initialized during app startup.
|
||||
"""
|
||||
if self._db_initialized:
|
||||
return
|
||||
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
|
||||
# Load pending items from database
|
||||
pending_items = await repository.get_pending_items()
|
||||
for item in pending_items:
|
||||
# Reset status if was downloading when saved
|
||||
if item.status == DownloadStatus.DOWNLOADING:
|
||||
item.status = DownloadStatus.PENDING
|
||||
await repository.update_status(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
self._add_to_pending_queue(item)
|
||||
|
||||
# Load failed items from database
|
||||
failed_items = await repository.get_failed_items()
|
||||
for item in failed_items:
|
||||
if item.retry_count < self._max_retries:
|
||||
item.status = DownloadStatus.PENDING
|
||||
await repository.update_status(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
self._add_to_pending_queue(item)
|
||||
else:
|
||||
self._failed_items.append(item)
|
||||
|
||||
# Load completed items for history
|
||||
completed_items = await repository.get_completed_items(limit=100)
|
||||
for item in completed_items:
|
||||
self._completed_items.append(item)
|
||||
|
||||
self._db_initialized = True
|
||||
|
||||
logger.info(
|
||||
"Queue restored from database",
|
||||
pending_count=len(self._pending_queue),
|
||||
failed_count=len(self._failed_items),
|
||||
completed_count=len(self._completed_items),
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error("Failed to load queue from database", error=str(e))
|
||||
# Continue without persistence - queue will work in memory only
|
||||
self._db_initialized = True
|
||||
|
||||
async def _save_to_database(self, item: DownloadItem) -> DownloadItem:
|
||||
"""Save or update an item in the database.
|
||||
|
||||
Args:
|
||||
item: Download item to save
|
||||
|
||||
Returns:
|
||||
Saved item with database ID
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.save_item(item)
|
||||
except Exception as e:
|
||||
logger.error("Failed to save item to database", error=str(e))
|
||||
return item
|
||||
|
||||
async def _update_status_in_database(
|
||||
self,
|
||||
item_id: str,
|
||||
status: DownloadStatus,
|
||||
error: Optional[str] = None,
|
||||
) -> bool:
|
||||
"""Update item status in the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
status: New status
|
||||
error: Optional error message
|
||||
|
||||
Returns:
|
||||
True if update succeeded
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.update_status(item_id, status, error)
|
||||
except Exception as e:
|
||||
logger.error("Failed to update status in database", error=str(e))
|
||||
return False
|
||||
|
||||
async def _update_progress_in_database(
|
||||
self,
|
||||
item_id: str,
|
||||
progress: float,
|
||||
downloaded: int,
|
||||
total: Optional[int],
|
||||
speed: Optional[float],
|
||||
) -> bool:
|
||||
"""Update download progress in the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
progress: Progress percentage
|
||||
downloaded: Downloaded bytes
|
||||
total: Total bytes
|
||||
speed: Download speed in bytes/sec
|
||||
|
||||
Returns:
|
||||
True if update succeeded
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.update_progress(
|
||||
item_id, progress, downloaded, total, speed
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error("Failed to update progress in database", error=str(e))
|
||||
return False
|
||||
|
||||
async def _delete_from_database(self, item_id: str) -> bool:
|
||||
"""Delete an item from the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
|
||||
Returns:
|
||||
True if delete succeeded
|
||||
"""
|
||||
try:
|
||||
repository = self._get_repository()
|
||||
return await repository.delete_item(item_id)
|
||||
except Exception as e:
|
||||
logger.error("Failed to delete from database", error=str(e))
|
||||
return False
|
||||
|
||||
async def _init_queue_progress(self) -> None:
|
||||
"""Initialize the download queue progress tracking.
|
||||
|
||||
@ -165,69 +317,6 @@ class DownloadService:
|
||||
"""Generate unique identifier for download items."""
|
||||
return str(uuid.uuid4())
|
||||
|
||||
def _load_queue(self) -> None:
|
||||
"""Load persisted queue from disk."""
|
||||
try:
|
||||
if self._persistence_path.exists():
|
||||
with open(self._persistence_path, "r", encoding="utf-8") as f:
|
||||
data = json.load(f)
|
||||
|
||||
# Restore pending items
|
||||
for item_dict in data.get("pending", []):
|
||||
item = DownloadItem(**item_dict)
|
||||
# Reset status if was downloading when saved
|
||||
if item.status == DownloadStatus.DOWNLOADING:
|
||||
item.status = DownloadStatus.PENDING
|
||||
self._add_to_pending_queue(item)
|
||||
|
||||
# Restore failed items that can be retried
|
||||
for item_dict in data.get("failed", []):
|
||||
item = DownloadItem(**item_dict)
|
||||
if item.retry_count < self._max_retries:
|
||||
item.status = DownloadStatus.PENDING
|
||||
self._add_to_pending_queue(item)
|
||||
else:
|
||||
self._failed_items.append(item)
|
||||
|
||||
logger.info(
|
||||
"Queue restored from disk",
|
||||
pending_count=len(self._pending_queue),
|
||||
failed_count=len(self._failed_items),
|
||||
)
|
||||
except Exception as e:
|
||||
logger.error("Failed to load persisted queue", error=str(e))
|
||||
|
||||
def _save_queue(self) -> None:
|
||||
"""Persist current queue state to disk."""
|
||||
try:
|
||||
self._persistence_path.parent.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
active_items = (
|
||||
[self._active_download] if self._active_download else []
|
||||
)
|
||||
|
||||
data = {
|
||||
"pending": [
|
||||
item.model_dump(mode="json")
|
||||
for item in self._pending_queue
|
||||
],
|
||||
"active": [
|
||||
item.model_dump(mode="json") for item in active_items
|
||||
],
|
||||
"failed": [
|
||||
item.model_dump(mode="json")
|
||||
for item in self._failed_items
|
||||
],
|
||||
"timestamp": datetime.now(timezone.utc).isoformat(),
|
||||
}
|
||||
|
||||
with open(self._persistence_path, "w", encoding="utf-8") as f:
|
||||
json.dump(data, f, indent=2)
|
||||
|
||||
logger.debug("Queue persisted to disk")
|
||||
except Exception as e:
|
||||
logger.error("Failed to persist queue", error=str(e))
|
||||
|
||||
async def add_to_queue(
|
||||
self,
|
||||
serie_id: str,
|
||||
@ -274,22 +363,23 @@ class DownloadService:
|
||||
added_at=datetime.now(timezone.utc),
|
||||
)
|
||||
|
||||
# Always append to end (FIFO order)
|
||||
self._add_to_pending_queue(item, front=False)
|
||||
# Save to database first to get persistent ID
|
||||
saved_item = await self._save_to_database(item)
|
||||
|
||||
created_ids.append(item.id)
|
||||
# Add to in-memory cache
|
||||
self._add_to_pending_queue(saved_item, front=False)
|
||||
|
||||
created_ids.append(saved_item.id)
|
||||
|
||||
logger.info(
|
||||
"Item added to queue",
|
||||
item_id=item.id,
|
||||
item_id=saved_item.id,
|
||||
serie_key=serie_id,
|
||||
serie_name=serie_name,
|
||||
season=episode.season,
|
||||
episode=episode.episode,
|
||||
)
|
||||
|
||||
self._save_queue()
|
||||
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@ -333,6 +423,10 @@ class DownloadService:
|
||||
item.completed_at = datetime.now(timezone.utc)
|
||||
self._failed_items.append(item)
|
||||
self._active_download = None
|
||||
# Update status in database
|
||||
await self._update_status_in_database(
|
||||
item_id, DownloadStatus.CANCELLED
|
||||
)
|
||||
removed_ids.append(item_id)
|
||||
logger.info("Cancelled active download", item_id=item_id)
|
||||
continue
|
||||
@ -342,13 +436,14 @@ class DownloadService:
|
||||
item = self._pending_items_by_id[item_id]
|
||||
self._pending_queue.remove(item)
|
||||
del self._pending_items_by_id[item_id]
|
||||
# Delete from database
|
||||
await self._delete_from_database(item_id)
|
||||
removed_ids.append(item_id)
|
||||
logger.info(
|
||||
"Removed from pending queue", item_id=item_id
|
||||
)
|
||||
|
||||
if removed_ids:
|
||||
self._save_queue()
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@ -379,6 +474,10 @@ class DownloadService:
|
||||
|
||||
Raises:
|
||||
DownloadServiceError: If reordering fails
|
||||
|
||||
Note:
|
||||
Reordering is done in-memory only. Database priority is not
|
||||
updated since the in-memory queue defines the actual order.
|
||||
"""
|
||||
try:
|
||||
# Build new queue based on specified order
|
||||
@ -399,9 +498,6 @@ class DownloadService:
|
||||
# Replace queue
|
||||
self._pending_queue = new_queue
|
||||
|
||||
# Save updated queue
|
||||
self._save_queue()
|
||||
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@ -692,13 +788,15 @@ class DownloadService:
|
||||
Number of items cleared
|
||||
"""
|
||||
count = len(self._pending_queue)
|
||||
|
||||
# Delete all pending items from database
|
||||
for item_id in list(self._pending_items_by_id.keys()):
|
||||
await self._delete_from_database(item_id)
|
||||
|
||||
self._pending_queue.clear()
|
||||
self._pending_items_by_id.clear()
|
||||
logger.info("Cleared pending items", count=count)
|
||||
|
||||
# Save queue state
|
||||
self._save_queue()
|
||||
|
||||
# Notify via progress service
|
||||
if count > 0:
|
||||
queue_status = await self.get_queue_status()
|
||||
@ -749,6 +847,11 @@ class DownloadService:
|
||||
self._add_to_pending_queue(item)
|
||||
retried_ids.append(item.id)
|
||||
|
||||
# Update status in database
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
|
||||
logger.info(
|
||||
"Retrying failed item",
|
||||
item_id=item.id,
|
||||
@ -756,7 +859,6 @@ class DownloadService:
|
||||
)
|
||||
|
||||
if retried_ids:
|
||||
self._save_queue()
|
||||
# Notify via progress service
|
||||
queue_status = await self.get_queue_status()
|
||||
await self._progress_service.update_progress(
|
||||
@ -790,10 +892,13 @@ class DownloadService:
|
||||
logger.info("Skipping download due to shutdown")
|
||||
return
|
||||
|
||||
# Update status
|
||||
# Update status in memory and database
|
||||
item.status = DownloadStatus.DOWNLOADING
|
||||
item.started_at = datetime.now(timezone.utc)
|
||||
self._active_download = item
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.DOWNLOADING
|
||||
)
|
||||
|
||||
logger.info(
|
||||
"Starting download",
|
||||
@ -809,7 +914,8 @@ class DownloadService:
|
||||
# - download started/progress/completed/failed events
|
||||
# - All updates forwarded to ProgressService
|
||||
# - ProgressService broadcasts to WebSocket clients
|
||||
# Use serie_folder for filesystem operations and serie_id (key) for identification
|
||||
# Use serie_folder for filesystem operations
|
||||
# and serie_id (key) for identification
|
||||
if not item.serie_folder:
|
||||
raise DownloadServiceError(
|
||||
f"Missing serie_folder for download item {item.id}. "
|
||||
@ -835,6 +941,11 @@ class DownloadService:
|
||||
|
||||
self._completed_items.append(item)
|
||||
|
||||
# Update database
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.COMPLETED
|
||||
)
|
||||
|
||||
logger.info(
|
||||
"Download completed successfully", item_id=item.id
|
||||
)
|
||||
@ -849,9 +960,15 @@ class DownloadService:
|
||||
)
|
||||
item.status = DownloadStatus.CANCELLED
|
||||
item.completed_at = datetime.now(timezone.utc)
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.CANCELLED
|
||||
)
|
||||
# Return item to pending queue if not shutting down
|
||||
if not self._is_shutting_down:
|
||||
self._add_to_pending_queue(item, front=True)
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.PENDING
|
||||
)
|
||||
raise # Re-raise to properly cancel the task
|
||||
|
||||
except Exception as e:
|
||||
@ -861,6 +978,11 @@ class DownloadService:
|
||||
item.error = str(e)
|
||||
self._failed_items.append(item)
|
||||
|
||||
# Update database with error
|
||||
await self._update_status_in_database(
|
||||
item.id, DownloadStatus.FAILED, str(e)
|
||||
)
|
||||
|
||||
logger.error(
|
||||
"Download failed",
|
||||
item_id=item.id,
|
||||
@ -874,8 +996,6 @@ class DownloadService:
|
||||
# Remove from active downloads
|
||||
if self._active_download and self._active_download.id == item.id:
|
||||
self._active_download = None
|
||||
|
||||
self._save_queue()
|
||||
|
||||
async def start(self) -> None:
|
||||
"""Initialize the download queue service (compatibility method).
|
||||
@ -896,17 +1016,15 @@ class DownloadService:
|
||||
self._is_stopped = True
|
||||
|
||||
# Cancel active download task if running
|
||||
if self._active_download_task and not self._active_download_task.done():
|
||||
active_task = self._active_download_task
|
||||
if active_task and not active_task.done():
|
||||
logger.info("Cancelling active download task...")
|
||||
self._active_download_task.cancel()
|
||||
active_task.cancel()
|
||||
try:
|
||||
await self._active_download_task
|
||||
await active_task
|
||||
except asyncio.CancelledError:
|
||||
logger.info("Active download task cancelled")
|
||||
|
||||
# Save final state
|
||||
self._save_queue()
|
||||
|
||||
# Shutdown executor immediately, don't wait for tasks
|
||||
logger.info("Shutting down thread pool executor...")
|
||||
self._executor.shutdown(wait=False, cancel_futures=True)
|
||||
|
||||
753
src/server/services/queue_repository.py
Normal file
753
src/server/services/queue_repository.py
Normal file
@ -0,0 +1,753 @@
|
||||
"""Queue repository adapter for database-backed download queue operations.
|
||||
|
||||
This module provides a repository adapter that wraps the DownloadQueueService
|
||||
and provides the interface needed by DownloadService for queue persistence.
|
||||
|
||||
The repository pattern abstracts the database operations from the business logic,
|
||||
allowing the DownloadService to work with domain models (DownloadItem) while
|
||||
the repository handles conversion to/from database models (DownloadQueueItem).
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
import logging
|
||||
from datetime import datetime, timezone
|
||||
from typing import Callable, List, Optional
|
||||
|
||||
from sqlalchemy.ext.asyncio import AsyncSession
|
||||
|
||||
from src.server.database.models import AnimeSeries
|
||||
from src.server.database.models import DownloadPriority as DBDownloadPriority
|
||||
from src.server.database.models import DownloadQueueItem as DBDownloadQueueItem
|
||||
from src.server.database.models import DownloadStatus as DBDownloadStatus
|
||||
from src.server.database.service import AnimeSeriesService, DownloadQueueService
|
||||
from src.server.models.download import (
|
||||
DownloadItem,
|
||||
DownloadPriority,
|
||||
DownloadProgress,
|
||||
DownloadStatus,
|
||||
EpisodeIdentifier,
|
||||
)
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
|
||||
class QueueRepositoryError(Exception):
|
||||
"""Repository-level exception for queue operations."""
|
||||
|
||||
|
||||
class QueueRepository:
|
||||
"""Repository adapter for database-backed download queue operations.
|
||||
|
||||
Provides clean interface for queue operations while handling
|
||||
model conversion between Pydantic (DownloadItem) and SQLAlchemy
|
||||
(DownloadQueueItem) models.
|
||||
|
||||
Attributes:
|
||||
_db_session_factory: Factory function to create database sessions
|
||||
"""
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
db_session_factory: Callable[[], AsyncSession],
|
||||
) -> None:
|
||||
"""Initialize the queue repository.
|
||||
|
||||
Args:
|
||||
db_session_factory: Factory function that returns AsyncSession instances
|
||||
"""
|
||||
self._db_session_factory = db_session_factory
|
||||
logger.info("QueueRepository initialized")
|
||||
|
||||
# =========================================================================
|
||||
# Model Conversion Methods
|
||||
# =========================================================================
|
||||
|
||||
def _status_to_db(self, status: DownloadStatus) -> DBDownloadStatus:
|
||||
"""Convert Pydantic DownloadStatus to SQLAlchemy DownloadStatus.
|
||||
|
||||
Args:
|
||||
status: Pydantic status enum
|
||||
|
||||
Returns:
|
||||
SQLAlchemy status enum
|
||||
"""
|
||||
return DBDownloadStatus(status.value)
|
||||
|
||||
def _status_from_db(self, status: DBDownloadStatus) -> DownloadStatus:
|
||||
"""Convert SQLAlchemy DownloadStatus to Pydantic DownloadStatus.
|
||||
|
||||
Args:
|
||||
status: SQLAlchemy status enum
|
||||
|
||||
Returns:
|
||||
Pydantic status enum
|
||||
"""
|
||||
return DownloadStatus(status.value)
|
||||
|
||||
def _priority_to_db(self, priority: DownloadPriority) -> DBDownloadPriority:
|
||||
"""Convert Pydantic DownloadPriority to SQLAlchemy DownloadPriority.
|
||||
|
||||
Args:
|
||||
priority: Pydantic priority enum
|
||||
|
||||
Returns:
|
||||
SQLAlchemy priority enum
|
||||
"""
|
||||
# Handle case differences (Pydantic uses uppercase, DB uses lowercase)
|
||||
return DBDownloadPriority(priority.value.lower())
|
||||
|
||||
def _priority_from_db(self, priority: DBDownloadPriority) -> DownloadPriority:
|
||||
"""Convert SQLAlchemy DownloadPriority to Pydantic DownloadPriority.
|
||||
|
||||
Args:
|
||||
priority: SQLAlchemy priority enum
|
||||
|
||||
Returns:
|
||||
Pydantic priority enum
|
||||
"""
|
||||
# Handle case differences (DB uses lowercase, Pydantic uses uppercase)
|
||||
return DownloadPriority(priority.value.upper())
|
||||
|
||||
def _to_db_model(
|
||||
self,
|
||||
item: DownloadItem,
|
||||
series_id: int,
|
||||
) -> DBDownloadQueueItem:
|
||||
"""Convert DownloadItem to database model.
|
||||
|
||||
Args:
|
||||
item: Pydantic download item
|
||||
series_id: Database series ID (foreign key)
|
||||
|
||||
Returns:
|
||||
SQLAlchemy download queue item model
|
||||
"""
|
||||
return DBDownloadQueueItem(
|
||||
series_id=series_id,
|
||||
season=item.episode.season,
|
||||
episode_number=item.episode.episode,
|
||||
status=self._status_to_db(item.status),
|
||||
priority=self._priority_to_db(item.priority),
|
||||
progress_percent=item.progress.percent if item.progress else 0.0,
|
||||
downloaded_bytes=int(
|
||||
item.progress.downloaded_mb * 1024 * 1024
|
||||
) if item.progress else 0,
|
||||
total_bytes=int(
|
||||
item.progress.total_mb * 1024 * 1024
|
||||
) if item.progress and item.progress.total_mb else None,
|
||||
download_speed=(
|
||||
item.progress.speed_mbps * 1024 * 1024
|
||||
) if item.progress and item.progress.speed_mbps else None,
|
||||
error_message=item.error,
|
||||
retry_count=item.retry_count,
|
||||
download_url=str(item.source_url) if item.source_url else None,
|
||||
started_at=item.started_at,
|
||||
completed_at=item.completed_at,
|
||||
)
|
||||
|
||||
def _from_db_model(
|
||||
self,
|
||||
db_item: DBDownloadQueueItem,
|
||||
item_id: Optional[str] = None,
|
||||
) -> DownloadItem:
|
||||
"""Convert database model to DownloadItem.
|
||||
|
||||
Args:
|
||||
db_item: SQLAlchemy download queue item
|
||||
item_id: Optional override for item ID (uses db ID if not provided)
|
||||
|
||||
Returns:
|
||||
Pydantic download item
|
||||
"""
|
||||
# Build progress object if there's progress data
|
||||
progress = None
|
||||
if db_item.progress_percent > 0 or db_item.downloaded_bytes > 0:
|
||||
progress = DownloadProgress(
|
||||
percent=db_item.progress_percent,
|
||||
downloaded_mb=db_item.downloaded_bytes / (1024 * 1024),
|
||||
total_mb=(
|
||||
db_item.total_bytes / (1024 * 1024)
|
||||
if db_item.total_bytes else None
|
||||
),
|
||||
speed_mbps=(
|
||||
db_item.download_speed / (1024 * 1024)
|
||||
if db_item.download_speed else None
|
||||
),
|
||||
)
|
||||
|
||||
return DownloadItem(
|
||||
id=item_id or str(db_item.id),
|
||||
serie_id=db_item.series.key if db_item.series else "",
|
||||
serie_folder=db_item.series.folder if db_item.series else "",
|
||||
serie_name=db_item.series.name if db_item.series else "",
|
||||
episode=EpisodeIdentifier(
|
||||
season=db_item.season,
|
||||
episode=db_item.episode_number,
|
||||
),
|
||||
status=self._status_from_db(db_item.status),
|
||||
priority=self._priority_from_db(db_item.priority),
|
||||
added_at=db_item.created_at or datetime.now(timezone.utc),
|
||||
started_at=db_item.started_at,
|
||||
completed_at=db_item.completed_at,
|
||||
progress=progress,
|
||||
error=db_item.error_message,
|
||||
retry_count=db_item.retry_count,
|
||||
source_url=db_item.download_url,
|
||||
)
|
||||
|
||||
# =========================================================================
|
||||
# CRUD Operations
|
||||
# =========================================================================
|
||||
|
||||
async def save_item(
|
||||
self,
|
||||
item: DownloadItem,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> DownloadItem:
|
||||
"""Save a download item to the database.
|
||||
|
||||
Creates a new record if the item doesn't exist in the database.
|
||||
|
||||
Args:
|
||||
item: Download item to save
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Saved download item with database ID
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If save operation fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
# Find series by key
|
||||
series = await AnimeSeriesService.get_by_key(session, item.serie_id)
|
||||
|
||||
if not series:
|
||||
# Create series if it doesn't exist
|
||||
series = await AnimeSeriesService.create(
|
||||
db=session,
|
||||
key=item.serie_id,
|
||||
name=item.serie_name,
|
||||
site="", # Will be updated later if needed
|
||||
folder=item.serie_folder,
|
||||
)
|
||||
logger.info(
|
||||
"Created new series for queue item",
|
||||
key=item.serie_id,
|
||||
name=item.serie_name,
|
||||
)
|
||||
|
||||
# Create queue item
|
||||
db_item = await DownloadQueueService.create(
|
||||
db=session,
|
||||
series_id=series.id,
|
||||
season=item.episode.season,
|
||||
episode_number=item.episode.episode,
|
||||
priority=self._priority_to_db(item.priority),
|
||||
download_url=str(item.source_url) if item.source_url else None,
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
# Update the item ID with the database ID
|
||||
item.id = str(db_item.id)
|
||||
|
||||
logger.debug(
|
||||
"Saved queue item to database",
|
||||
item_id=item.id,
|
||||
serie_key=item.serie_id,
|
||||
)
|
||||
|
||||
return item
|
||||
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to save queue item", error=str(e))
|
||||
raise QueueRepositoryError(f"Failed to save item: {str(e)}") from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_item(
|
||||
self,
|
||||
item_id: str,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> Optional[DownloadItem]:
|
||||
"""Get a download item by ID.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID (database ID as string)
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Download item or None if not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_item = await DownloadQueueService.get_by_id(
|
||||
session, int(item_id)
|
||||
)
|
||||
|
||||
if not db_item:
|
||||
return None
|
||||
|
||||
return self._from_db_model(db_item, item_id)
|
||||
|
||||
except ValueError:
|
||||
# Invalid ID format
|
||||
return None
|
||||
except Exception as e:
|
||||
logger.error("Failed to get queue item", error=str(e))
|
||||
raise QueueRepositoryError(f"Failed to get item: {str(e)}") from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_pending_items(
|
||||
self,
|
||||
limit: Optional[int] = None,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get pending download items ordered by priority.
|
||||
|
||||
Args:
|
||||
limit: Optional maximum number of items to return
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of pending download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_pending(session, limit)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get pending items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get pending items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_active_item(
|
||||
self,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> Optional[DownloadItem]:
|
||||
"""Get the currently active (downloading) item.
|
||||
|
||||
Args:
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Active download item or None if none active
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_active(session)
|
||||
|
||||
if not db_items:
|
||||
return None
|
||||
|
||||
# Return first active item (should only be one)
|
||||
return self._from_db_model(db_items[0])
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get active item", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get active item: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_completed_items(
|
||||
self,
|
||||
limit: int = 100,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get completed download items.
|
||||
|
||||
Args:
|
||||
limit: Maximum number of items to return
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of completed download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_by_status(
|
||||
session, DBDownloadStatus.COMPLETED, limit
|
||||
)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get completed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get completed items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_failed_items(
|
||||
self,
|
||||
limit: int = 50,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get failed download items.
|
||||
|
||||
Args:
|
||||
limit: Maximum number of items to return
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of failed download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_by_status(
|
||||
session, DBDownloadStatus.FAILED, limit
|
||||
)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get failed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get failed items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def update_status(
|
||||
self,
|
||||
item_id: str,
|
||||
status: DownloadStatus,
|
||||
error: Optional[str] = None,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> bool:
|
||||
"""Update the status of a download item.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
status: New download status
|
||||
error: Optional error message for failed status
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
True if update succeeded, False if item not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If update fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
result = await DownloadQueueService.update_status(
|
||||
session,
|
||||
int(item_id),
|
||||
self._status_to_db(status),
|
||||
error,
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
success = result is not None
|
||||
|
||||
if success:
|
||||
logger.debug(
|
||||
"Updated queue item status",
|
||||
item_id=item_id,
|
||||
status=status.value,
|
||||
)
|
||||
|
||||
return success
|
||||
|
||||
except ValueError:
|
||||
return False
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to update status", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to update status: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def update_progress(
|
||||
self,
|
||||
item_id: str,
|
||||
progress: float,
|
||||
downloaded: int,
|
||||
total: Optional[int],
|
||||
speed: Optional[float],
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> bool:
|
||||
"""Update download progress for an item.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
progress: Progress percentage (0-100)
|
||||
downloaded: Downloaded bytes
|
||||
total: Total bytes (optional)
|
||||
speed: Download speed in bytes/second (optional)
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
True if update succeeded, False if item not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If update fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
result = await DownloadQueueService.update_progress(
|
||||
session,
|
||||
int(item_id),
|
||||
progress,
|
||||
downloaded,
|
||||
total,
|
||||
speed,
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
return result is not None
|
||||
|
||||
except ValueError:
|
||||
return False
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to update progress", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to update progress: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def delete_item(
|
||||
self,
|
||||
item_id: str,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> bool:
|
||||
"""Delete a download item from the database.
|
||||
|
||||
Args:
|
||||
item_id: Download item ID
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
True if item was deleted, False if not found
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If delete fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
result = await DownloadQueueService.delete(session, int(item_id))
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
if result:
|
||||
logger.debug("Deleted queue item", item_id=item_id)
|
||||
|
||||
return result
|
||||
|
||||
except ValueError:
|
||||
return False
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to delete item", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to delete item: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def clear_completed(
|
||||
self,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> int:
|
||||
"""Clear all completed download items.
|
||||
|
||||
Args:
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
Number of items cleared
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If operation fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
count = await DownloadQueueService.clear_completed(session)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
logger.info("Cleared completed items from queue", count=count)
|
||||
return count
|
||||
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to clear completed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to clear completed: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def get_all_items(
|
||||
self,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Get all download items regardless of status.
|
||||
|
||||
Args:
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of all download items
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If query fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.get_all(
|
||||
session, with_series=True
|
||||
)
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
logger.error("Failed to get all items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to get all items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
async def retry_failed_items(
|
||||
self,
|
||||
max_retries: int = 3,
|
||||
db: Optional[AsyncSession] = None,
|
||||
) -> List[DownloadItem]:
|
||||
"""Retry failed downloads that haven't exceeded max retries.
|
||||
|
||||
Args:
|
||||
max_retries: Maximum number of retry attempts
|
||||
db: Optional existing database session
|
||||
|
||||
Returns:
|
||||
List of items marked for retry
|
||||
|
||||
Raises:
|
||||
QueueRepositoryError: If operation fails
|
||||
"""
|
||||
session = db or self._db_session_factory()
|
||||
manage_session = db is None
|
||||
|
||||
try:
|
||||
db_items = await DownloadQueueService.retry_failed(
|
||||
session, max_retries
|
||||
)
|
||||
|
||||
if manage_session:
|
||||
await session.commit()
|
||||
|
||||
return [self._from_db_model(item) for item in db_items]
|
||||
|
||||
except Exception as e:
|
||||
if manage_session:
|
||||
await session.rollback()
|
||||
logger.error("Failed to retry failed items", error=str(e))
|
||||
raise QueueRepositoryError(
|
||||
f"Failed to retry failed items: {str(e)}"
|
||||
) from e
|
||||
finally:
|
||||
if manage_session:
|
||||
await session.close()
|
||||
|
||||
|
||||
# Singleton instance
|
||||
_queue_repository_instance: Optional[QueueRepository] = None
|
||||
|
||||
|
||||
def get_queue_repository(
|
||||
db_session_factory: Optional[Callable[[], AsyncSession]] = None,
|
||||
) -> QueueRepository:
|
||||
"""Get or create the QueueRepository singleton.
|
||||
|
||||
Args:
|
||||
db_session_factory: Optional factory function for database sessions.
|
||||
If not provided, uses default from connection module.
|
||||
|
||||
Returns:
|
||||
QueueRepository singleton instance
|
||||
"""
|
||||
global _queue_repository_instance
|
||||
|
||||
if _queue_repository_instance is None:
|
||||
if db_session_factory is None:
|
||||
# Use default session factory
|
||||
from src.server.database.connection import get_async_session_factory
|
||||
db_session_factory = get_async_session_factory
|
||||
|
||||
_queue_repository_instance = QueueRepository(db_session_factory)
|
||||
|
||||
return _queue_repository_instance
|
||||
@ -72,11 +72,14 @@ async def anime_service(mock_series_app, progress_service):
|
||||
|
||||
@pytest.fixture
|
||||
async def download_service(anime_service, progress_service):
|
||||
"""Create a DownloadService."""
|
||||
"""Create a DownloadService with mock queue repository."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=anime_service,
|
||||
progress_service=progress_service,
|
||||
persistence_path="/tmp/test_integration_progress_queue.json",
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
yield service
|
||||
await service.stop()
|
||||
|
||||
@ -88,9 +88,10 @@ def progress_service():
|
||||
|
||||
@pytest.fixture
|
||||
async def download_service(mock_series_app, progress_service, tmp_path):
|
||||
"""Create a DownloadService with dependencies."""
|
||||
import uuid
|
||||
persistence_path = tmp_path / f"test_queue_{uuid.uuid4()}.json"
|
||||
"""Create a DownloadService with mock repository for testing."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
|
||||
mock_repo = MockQueueRepository()
|
||||
|
||||
anime_service = AnimeService(
|
||||
series_app=mock_series_app,
|
||||
@ -101,7 +102,7 @@ async def download_service(mock_series_app, progress_service, tmp_path):
|
||||
service = DownloadService(
|
||||
anime_service=anime_service,
|
||||
progress_service=progress_service,
|
||||
persistence_path=str(persistence_path),
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
yield service
|
||||
await service.stop()
|
||||
@ -319,8 +320,6 @@ class TestServiceIdentifierConsistency:
|
||||
- Persisted data contains serie_id (key)
|
||||
- Data can be restored with correct identifiers
|
||||
"""
|
||||
import json
|
||||
|
||||
# Add item to queue
|
||||
await download_service.add_to_queue(
|
||||
serie_id="jujutsu-kaisen",
|
||||
@ -330,18 +329,13 @@ class TestServiceIdentifierConsistency:
|
||||
priority=DownloadPriority.NORMAL,
|
||||
)
|
||||
|
||||
# Read persisted data
|
||||
persistence_path = download_service._persistence_path
|
||||
with open(persistence_path, "r") as f:
|
||||
data = json.load(f)
|
||||
# Verify item is in pending queue (in-memory cache synced with DB)
|
||||
pending_items = list(download_service._pending_queue)
|
||||
assert len(pending_items) == 1
|
||||
|
||||
# Verify persisted data structure
|
||||
assert "pending" in data
|
||||
assert len(data["pending"]) == 1
|
||||
|
||||
persisted_item = data["pending"][0]
|
||||
assert persisted_item["serie_id"] == "jujutsu-kaisen"
|
||||
assert persisted_item["serie_folder"] == "Jujutsu Kaisen (2020)"
|
||||
persisted_item = pending_items[0]
|
||||
assert persisted_item.serie_id == "jujutsu-kaisen"
|
||||
assert persisted_item.serie_folder == "Jujutsu Kaisen (2020)"
|
||||
|
||||
|
||||
class TestWebSocketIdentifierConsistency:
|
||||
|
||||
@ -69,16 +69,17 @@ async def anime_service(mock_series_app, progress_service):
|
||||
|
||||
@pytest.fixture
|
||||
async def download_service(anime_service, progress_service, tmp_path):
|
||||
"""Create a DownloadService with dependencies.
|
||||
"""Create a DownloadService with mock repository for testing.
|
||||
|
||||
Uses tmp_path to ensure each test has isolated queue storage.
|
||||
Uses mock repository to ensure each test has isolated queue storage.
|
||||
"""
|
||||
import uuid
|
||||
persistence_path = tmp_path / f"test_queue_{uuid.uuid4()}.json"
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=anime_service,
|
||||
progress_service=progress_service,
|
||||
persistence_path=str(persistence_path),
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
yield service, progress_service
|
||||
await service.stop()
|
||||
|
||||
@ -28,12 +28,13 @@ class TestDownloadQueueStress:
|
||||
|
||||
@pytest.fixture
|
||||
def download_service(self, mock_anime_service, tmp_path):
|
||||
"""Create download service with mock."""
|
||||
persistence_path = str(tmp_path / "test_queue.json")
|
||||
"""Create download service with mock repository."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
max_retries=3,
|
||||
persistence_path=persistence_path,
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
return service
|
||||
|
||||
@ -176,12 +177,13 @@ class TestDownloadMemoryUsage:
|
||||
|
||||
@pytest.fixture
|
||||
def download_service(self, mock_anime_service, tmp_path):
|
||||
"""Create download service with mock."""
|
||||
persistence_path = str(tmp_path / "test_queue.json")
|
||||
"""Create download service with mock repository."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
max_retries=3,
|
||||
persistence_path=persistence_path,
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
return service
|
||||
|
||||
@ -232,12 +234,13 @@ class TestDownloadConcurrency:
|
||||
|
||||
@pytest.fixture
|
||||
def download_service(self, mock_anime_service, tmp_path):
|
||||
"""Create download service with mock."""
|
||||
persistence_path = str(tmp_path / "test_queue.json")
|
||||
"""Create download service with mock repository."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
max_retries=3,
|
||||
persistence_path=persistence_path,
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
return service
|
||||
|
||||
@ -321,11 +324,12 @@ class TestDownloadErrorHandling:
|
||||
self, mock_failing_anime_service, tmp_path
|
||||
):
|
||||
"""Create download service with failing mock."""
|
||||
persistence_path = str(tmp_path / "test_queue.json")
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=mock_failing_anime_service,
|
||||
max_retries=3,
|
||||
persistence_path=persistence_path,
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
return service
|
||||
|
||||
@ -338,12 +342,13 @@ class TestDownloadErrorHandling:
|
||||
|
||||
@pytest.fixture
|
||||
def download_service(self, mock_anime_service, tmp_path):
|
||||
"""Create download service with mock."""
|
||||
persistence_path = str(tmp_path / "test_queue.json")
|
||||
"""Create download service with mock repository."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
mock_repo = MockQueueRepository()
|
||||
service = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
max_retries=3,
|
||||
persistence_path=persistence_path,
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
return service
|
||||
|
||||
|
||||
@ -102,27 +102,20 @@ async def anime_service(mock_series_app, progress_service):
|
||||
|
||||
@pytest.fixture
|
||||
async def download_service(anime_service, progress_service):
|
||||
"""Create a DownloadService with dependencies."""
|
||||
import os
|
||||
persistence_path = "/tmp/test_download_progress_queue.json"
|
||||
"""Create a DownloadService with mock repository for testing."""
|
||||
from tests.unit.test_download_service import MockQueueRepository
|
||||
|
||||
# Remove any existing queue file
|
||||
if os.path.exists(persistence_path):
|
||||
os.remove(persistence_path)
|
||||
mock_repo = MockQueueRepository()
|
||||
|
||||
service = DownloadService(
|
||||
anime_service=anime_service,
|
||||
progress_service=progress_service,
|
||||
persistence_path=persistence_path,
|
||||
queue_repository=mock_repo,
|
||||
)
|
||||
|
||||
yield service, progress_service
|
||||
|
||||
await service.stop()
|
||||
|
||||
# Clean up after test
|
||||
if os.path.exists(persistence_path):
|
||||
os.remove(persistence_path)
|
||||
|
||||
|
||||
class TestDownloadProgressWebSocket:
|
||||
|
||||
@ -1,14 +1,13 @@
|
||||
"""Unit tests for the download queue service.
|
||||
|
||||
Tests cover queue management, manual download control, persistence,
|
||||
Tests cover queue management, manual download control, database persistence,
|
||||
and error scenarios for the simplified download service.
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
import asyncio
|
||||
import json
|
||||
from datetime import datetime, timezone
|
||||
from pathlib import Path
|
||||
from typing import Dict, List, Optional
|
||||
from unittest.mock import AsyncMock, MagicMock
|
||||
|
||||
import pytest
|
||||
@ -20,7 +19,125 @@ from src.server.models.download import (
|
||||
EpisodeIdentifier,
|
||||
)
|
||||
from src.server.services.anime_service import AnimeService
|
||||
from src.server.services.download_service import DownloadService, DownloadServiceError
|
||||
from src.server.services.download_service import (
|
||||
DownloadService,
|
||||
DownloadServiceError,
|
||||
)
|
||||
|
||||
|
||||
class MockQueueRepository:
|
||||
"""Mock implementation of QueueRepository for testing.
|
||||
|
||||
This provides an in-memory storage that mimics the database repository
|
||||
behavior without requiring actual database connections.
|
||||
"""
|
||||
|
||||
def __init__(self):
|
||||
"""Initialize mock repository with in-memory storage."""
|
||||
self._items: Dict[str, DownloadItem] = {}
|
||||
|
||||
async def save_item(self, item: DownloadItem) -> DownloadItem:
|
||||
"""Save item to in-memory storage."""
|
||||
self._items[item.id] = item
|
||||
return item
|
||||
|
||||
async def get_item(self, item_id: str) -> Optional[DownloadItem]:
|
||||
"""Get item by ID from in-memory storage."""
|
||||
return self._items.get(item_id)
|
||||
|
||||
async def get_pending_items(self) -> List[DownloadItem]:
|
||||
"""Get all pending items."""
|
||||
return [
|
||||
item for item in self._items.values()
|
||||
if item.status == DownloadStatus.PENDING
|
||||
]
|
||||
|
||||
async def get_active_item(self) -> Optional[DownloadItem]:
|
||||
"""Get the currently active item."""
|
||||
for item in self._items.values():
|
||||
if item.status == DownloadStatus.DOWNLOADING:
|
||||
return item
|
||||
return None
|
||||
|
||||
async def get_completed_items(
|
||||
self, limit: int = 100
|
||||
) -> List[DownloadItem]:
|
||||
"""Get completed items."""
|
||||
completed = [
|
||||
item for item in self._items.values()
|
||||
if item.status == DownloadStatus.COMPLETED
|
||||
]
|
||||
return completed[:limit]
|
||||
|
||||
async def get_failed_items(self, limit: int = 50) -> List[DownloadItem]:
|
||||
"""Get failed items."""
|
||||
failed = [
|
||||
item for item in self._items.values()
|
||||
if item.status == DownloadStatus.FAILED
|
||||
]
|
||||
return failed[:limit]
|
||||
|
||||
async def update_status(
|
||||
self,
|
||||
item_id: str,
|
||||
status: DownloadStatus,
|
||||
error: Optional[str] = None
|
||||
) -> bool:
|
||||
"""Update item status."""
|
||||
if item_id not in self._items:
|
||||
return False
|
||||
self._items[item_id].status = status
|
||||
if error:
|
||||
self._items[item_id].error = error
|
||||
if status == DownloadStatus.COMPLETED:
|
||||
self._items[item_id].completed_at = datetime.now(timezone.utc)
|
||||
elif status == DownloadStatus.DOWNLOADING:
|
||||
self._items[item_id].started_at = datetime.now(timezone.utc)
|
||||
return True
|
||||
|
||||
async def update_progress(
|
||||
self,
|
||||
item_id: str,
|
||||
progress: float,
|
||||
downloaded: int,
|
||||
total: int,
|
||||
speed: float
|
||||
) -> bool:
|
||||
"""Update download progress."""
|
||||
if item_id not in self._items:
|
||||
return False
|
||||
item = self._items[item_id]
|
||||
if item.progress is None:
|
||||
from src.server.models.download import DownloadProgress
|
||||
item.progress = DownloadProgress(
|
||||
percent=progress,
|
||||
downloaded_bytes=downloaded,
|
||||
total_bytes=total,
|
||||
speed_bps=speed
|
||||
)
|
||||
else:
|
||||
item.progress.percent = progress
|
||||
item.progress.downloaded_bytes = downloaded
|
||||
item.progress.total_bytes = total
|
||||
item.progress.speed_bps = speed
|
||||
return True
|
||||
|
||||
async def delete_item(self, item_id: str) -> bool:
|
||||
"""Delete item from storage."""
|
||||
if item_id in self._items:
|
||||
del self._items[item_id]
|
||||
return True
|
||||
return False
|
||||
|
||||
async def clear_completed(self) -> int:
|
||||
"""Clear all completed items."""
|
||||
completed_ids = [
|
||||
item_id for item_id, item in self._items.items()
|
||||
if item.status == DownloadStatus.COMPLETED
|
||||
]
|
||||
for item_id in completed_ids:
|
||||
del self._items[item_id]
|
||||
return len(completed_ids)
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
@ -32,18 +149,18 @@ def mock_anime_service():
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def temp_persistence_path(tmp_path):
|
||||
"""Create a temporary persistence path."""
|
||||
return str(tmp_path / "test_queue.json")
|
||||
def mock_queue_repository():
|
||||
"""Create a mock QueueRepository for testing."""
|
||||
return MockQueueRepository()
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def download_service(mock_anime_service, temp_persistence_path):
|
||||
def download_service(mock_anime_service, mock_queue_repository):
|
||||
"""Create a DownloadService instance for testing."""
|
||||
return DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
queue_repository=mock_queue_repository,
|
||||
max_retries=3,
|
||||
persistence_path=temp_persistence_path,
|
||||
)
|
||||
|
||||
|
||||
@ -51,12 +168,12 @@ class TestDownloadServiceInitialization:
|
||||
"""Test download service initialization."""
|
||||
|
||||
def test_initialization_creates_queues(
|
||||
self, mock_anime_service, temp_persistence_path
|
||||
self, mock_anime_service, mock_queue_repository
|
||||
):
|
||||
"""Test that initialization creates empty queues."""
|
||||
service = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
persistence_path=temp_persistence_path,
|
||||
queue_repository=mock_queue_repository,
|
||||
)
|
||||
|
||||
assert len(service._pending_queue) == 0
|
||||
@ -65,45 +182,30 @@ class TestDownloadServiceInitialization:
|
||||
assert len(service._failed_items) == 0
|
||||
assert service._is_stopped is True
|
||||
|
||||
def test_initialization_loads_persisted_queue(
|
||||
self, mock_anime_service, temp_persistence_path
|
||||
@pytest.mark.asyncio
|
||||
async def test_initialization_loads_persisted_queue(
|
||||
self, mock_anime_service, mock_queue_repository
|
||||
):
|
||||
"""Test that initialization loads persisted queue state."""
|
||||
# Create a persisted queue file
|
||||
persistence_file = Path(temp_persistence_path)
|
||||
persistence_file.parent.mkdir(parents=True, exist_ok=True)
|
||||
|
||||
test_data = {
|
||||
"pending": [
|
||||
{
|
||||
"id": "test-id-1",
|
||||
"serie_id": "series-1",
|
||||
"serie_folder": "test-series", # Added missing field
|
||||
"serie_name": "Test Series",
|
||||
"episode": {"season": 1, "episode": 1, "title": None},
|
||||
"status": "pending",
|
||||
"priority": "NORMAL", # Must be uppercase
|
||||
"added_at": datetime.now(timezone.utc).isoformat(),
|
||||
"started_at": None,
|
||||
"completed_at": None,
|
||||
"progress": None,
|
||||
"error": None,
|
||||
"retry_count": 0,
|
||||
"source_url": None,
|
||||
}
|
||||
],
|
||||
"active": [],
|
||||
"failed": [],
|
||||
"timestamp": datetime.now(timezone.utc).isoformat(),
|
||||
}
|
||||
|
||||
with open(persistence_file, "w", encoding="utf-8") as f:
|
||||
json.dump(test_data, f)
|
||||
"""Test that initialization loads persisted queue from database."""
|
||||
# Pre-populate the mock repository with a pending item
|
||||
test_item = DownloadItem(
|
||||
id="test-id-1",
|
||||
serie_id="series-1",
|
||||
serie_folder="test-series",
|
||||
serie_name="Test Series",
|
||||
episode=EpisodeIdentifier(season=1, episode=1),
|
||||
status=DownloadStatus.PENDING,
|
||||
priority=DownloadPriority.NORMAL,
|
||||
added_at=datetime.now(timezone.utc),
|
||||
)
|
||||
await mock_queue_repository.save_item(test_item)
|
||||
|
||||
# Create service and initialize from database
|
||||
service = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
persistence_path=temp_persistence_path,
|
||||
queue_repository=mock_queue_repository,
|
||||
)
|
||||
await service.initialize()
|
||||
|
||||
assert len(service._pending_queue) == 1
|
||||
assert service._pending_queue[0].id == "test-id-1"
|
||||
@ -391,11 +493,13 @@ class TestQueueControl:
|
||||
|
||||
|
||||
class TestPersistence:
|
||||
"""Test queue persistence functionality."""
|
||||
"""Test queue persistence functionality with database backend."""
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_queue_persistence(self, download_service):
|
||||
"""Test that queue state is persisted to disk."""
|
||||
async def test_queue_persistence(
|
||||
self, download_service, mock_queue_repository
|
||||
):
|
||||
"""Test that queue state is persisted to database."""
|
||||
await download_service.add_to_queue(
|
||||
serie_id="series-1",
|
||||
serie_folder="series",
|
||||
@ -403,26 +507,20 @@ class TestPersistence:
|
||||
episodes=[EpisodeIdentifier(season=1, episode=1)],
|
||||
)
|
||||
|
||||
# Persistence file should exist
|
||||
persistence_path = Path(download_service._persistence_path)
|
||||
assert persistence_path.exists()
|
||||
|
||||
# Check file contents
|
||||
with open(persistence_path, "r") as f:
|
||||
data = json.load(f)
|
||||
|
||||
assert len(data["pending"]) == 1
|
||||
assert data["pending"][0]["serie_id"] == "series-1"
|
||||
# Item should be saved in mock repository
|
||||
pending_items = await mock_queue_repository.get_pending_items()
|
||||
assert len(pending_items) == 1
|
||||
assert pending_items[0].serie_id == "series-1"
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_queue_recovery_after_restart(
|
||||
self, mock_anime_service, temp_persistence_path
|
||||
self, mock_anime_service, mock_queue_repository
|
||||
):
|
||||
"""Test that queue is recovered after service restart."""
|
||||
# Create and populate first service
|
||||
service1 = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
persistence_path=temp_persistence_path,
|
||||
queue_repository=mock_queue_repository,
|
||||
)
|
||||
|
||||
await service1.add_to_queue(
|
||||
@ -435,11 +533,13 @@ class TestPersistence:
|
||||
],
|
||||
)
|
||||
|
||||
# Create new service with same persistence path
|
||||
# Create new service with same repository (simulating restart)
|
||||
service2 = DownloadService(
|
||||
anime_service=mock_anime_service,
|
||||
persistence_path=temp_persistence_path,
|
||||
queue_repository=mock_queue_repository,
|
||||
)
|
||||
# Initialize to load from database to recover state
|
||||
await service2.initialize()
|
||||
|
||||
# Should recover pending items
|
||||
assert len(service2._pending_queue) == 2
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user