Replaces per-member Nextcloud credentials with a BandStorage model that supports multiple providers. Credentials are Fernet-encrypted at rest; worker receives audio via an internal streaming endpoint instead of direct storage access. - Add BandStorage DB model with partial unique index (one active per band) - Add migrations 0007 (create band_storage) and 0008 (drop old nc columns) - Add StorageFactory that builds the correct StorageClient from BandStorage - Add storage router: connect/nextcloud, OAuth2 authorize/callback, list, disconnect - Add Fernet encryption helpers in security/encryption.py - Rewrite watcher for per-band polling via internal API config endpoint - Update worker to stream audio from API instead of accessing storage directly - Update frontend: new storage API in bands.ts, rewritten StorageSection, simplified band creation modal (no storage step) - Add STORAGE_ENCRYPTION_KEY to all docker-compose files Co-Authored-By: Claude Sonnet 4.6 <noreply@anthropic.com>
191 lines
7.5 KiB
Python
Executable File
191 lines
7.5 KiB
Python
Executable File
"""Storage scan logic: walk a band's storage folder and import audio files.
|
|
|
|
Works against any ``StorageClient`` implementation — Nextcloud, Google Drive, etc.
|
|
``StorageClient.list_folder`` must return ``FileMetadata`` objects whose ``path``
|
|
field is a *provider-relative* path (i.e. the DAV prefix has already been stripped
|
|
by the client implementation).
|
|
"""
|
|
|
|
from __future__ import annotations
|
|
|
|
import logging
|
|
from collections.abc import AsyncGenerator
|
|
from pathlib import Path
|
|
|
|
from sqlalchemy.ext.asyncio import AsyncSession
|
|
|
|
from rehearsalhub.repositories.audio_version import AudioVersionRepository
|
|
from rehearsalhub.repositories.rehearsal_session import RehearsalSessionRepository
|
|
from rehearsalhub.repositories.song import SongRepository
|
|
from rehearsalhub.schemas.audio_version import AudioVersionCreate
|
|
from rehearsalhub.schemas.song import SongRead
|
|
from rehearsalhub.services.session import extract_session_folder, parse_rehearsal_date
|
|
from rehearsalhub.services.song import SongService
|
|
from rehearsalhub.storage.protocol import StorageClient
|
|
|
|
log = logging.getLogger(__name__)
|
|
|
|
AUDIO_EXTENSIONS = {".mp3", ".wav", ".flac", ".ogg", ".m4a", ".aac", ".opus"}
|
|
|
|
# Maximum folder depth to recurse into below the band root.
|
|
# Depth 0 = band root, 1 = YYMMDD folder, 2 = song subfolder, 3 = safety margin.
|
|
MAX_SCAN_DEPTH = 3
|
|
|
|
|
|
async def collect_audio_files(
|
|
storage: StorageClient,
|
|
folder_path: str,
|
|
max_depth: int = MAX_SCAN_DEPTH,
|
|
_depth: int = 0,
|
|
) -> AsyncGenerator[str, None]:
|
|
"""Recursively yield provider-relative audio file paths under *folder_path*.
|
|
|
|
``storage.list_folder`` is expected to return ``FileMetadata`` with paths
|
|
already normalised to provider-relative form (no host, no DAV prefix).
|
|
"""
|
|
if _depth > max_depth:
|
|
log.debug("Max depth %d exceeded at '%s', stopping recursion", max_depth, folder_path)
|
|
return
|
|
|
|
try:
|
|
items = await storage.list_folder(folder_path)
|
|
except Exception as exc:
|
|
log.warning("Could not list folder '%s': %s", folder_path, exc)
|
|
return
|
|
|
|
log.info("scan depth=%d folder='%s' entries=%d", _depth, folder_path, len(items))
|
|
|
|
for item in items:
|
|
path = item.path.lstrip("/")
|
|
if path.endswith("/"):
|
|
log.info(" → subdir: %s", path)
|
|
async for subpath in collect_audio_files(storage, path, max_depth, _depth + 1):
|
|
yield subpath
|
|
else:
|
|
ext = Path(path).suffix.lower()
|
|
if ext in AUDIO_EXTENSIONS:
|
|
log.info(" → audio file: %s", path)
|
|
yield path
|
|
elif ext:
|
|
log.debug(" → skip (ext=%s): %s", ext, path)
|
|
|
|
|
|
async def scan_band_folder(
|
|
db_session: AsyncSession,
|
|
storage: StorageClient,
|
|
band_id,
|
|
band_folder: str,
|
|
member_id,
|
|
) -> AsyncGenerator[dict, None]:
|
|
"""Async generator that scans *band_folder* and yields event dicts:
|
|
|
|
{"type": "progress", "message": str}
|
|
{"type": "song", "song": SongRead-dict, "is_new": bool}
|
|
{"type": "session", "session": {id, date, label}}
|
|
{"type": "skipped", "path": str, "reason": str}
|
|
{"type": "done", "stats": {found, imported, skipped}}
|
|
{"type": "error", "message": str}
|
|
"""
|
|
session_repo = RehearsalSessionRepository(db_session)
|
|
song_repo = SongRepository(db_session)
|
|
version_repo = AudioVersionRepository(db_session)
|
|
song_svc = SongService(db_session)
|
|
|
|
found = 0
|
|
imported = 0
|
|
skipped = 0
|
|
|
|
yield {"type": "progress", "message": f"Scanning {band_folder}…"}
|
|
|
|
async for nc_file_path in collect_audio_files(storage, band_folder):
|
|
found += 1
|
|
song_folder = str(Path(nc_file_path).parent).rstrip("/") + "/"
|
|
song_title = Path(nc_file_path).stem
|
|
|
|
# If the file sits directly inside a dated session folder (YYMMDD/file.wav),
|
|
# give it a unique virtual folder so each file becomes its own song.
|
|
session_folder_path = extract_session_folder(nc_file_path)
|
|
if session_folder_path and session_folder_path.rstrip("/") == song_folder.rstrip("/"):
|
|
song_folder = song_folder + song_title + "/"
|
|
|
|
yield {"type": "progress", "message": f"Checking {Path(nc_file_path).name}…"}
|
|
|
|
existing = await version_repo.get_by_nc_file_path(nc_file_path)
|
|
if existing is not None:
|
|
log.debug("scan: skipping already-registered '%s' (version %s)", nc_file_path, existing.id)
|
|
skipped += 1
|
|
yield {"type": "skipped", "path": nc_file_path, "reason": "already imported"}
|
|
continue
|
|
|
|
try:
|
|
meta = await storage.get_file_metadata(nc_file_path)
|
|
etag = meta.etag
|
|
except Exception as exc:
|
|
log.error("Metadata fetch failed for '%s': %s", nc_file_path, exc, exc_info=True)
|
|
skipped += 1
|
|
yield {"type": "skipped", "path": nc_file_path, "reason": f"metadata error: {exc}"}
|
|
continue
|
|
|
|
try:
|
|
rehearsal_date = parse_rehearsal_date(nc_file_path)
|
|
rehearsal_session_id = None
|
|
if rehearsal_date:
|
|
session_folder = extract_session_folder(nc_file_path) or song_folder
|
|
rs = await session_repo.get_or_create(band_id, rehearsal_date, session_folder)
|
|
rehearsal_session_id = rs.id
|
|
yield {
|
|
"type": "session",
|
|
"session": {
|
|
"id": str(rs.id),
|
|
"date": rs.date.isoformat(),
|
|
"label": rs.label,
|
|
"nc_folder_path": rs.nc_folder_path,
|
|
},
|
|
}
|
|
|
|
song = await song_repo.get_by_nc_folder_path(song_folder)
|
|
if song is None:
|
|
song = await song_repo.get_by_title_and_band(band_id, song_title)
|
|
is_new = song is None
|
|
if is_new:
|
|
log.info("Creating song '%s' folder='%s'", song_title, song_folder)
|
|
song = await song_repo.create(
|
|
band_id=band_id,
|
|
session_id=rehearsal_session_id,
|
|
title=song_title,
|
|
status="jam",
|
|
notes=None,
|
|
nc_folder_path=song_folder,
|
|
created_by=member_id,
|
|
)
|
|
elif rehearsal_session_id and song.session_id is None:
|
|
song = await song_repo.update(song, session_id=rehearsal_session_id)
|
|
|
|
version = await song_svc.register_version(
|
|
song.id,
|
|
AudioVersionCreate(
|
|
nc_file_path=nc_file_path,
|
|
nc_file_etag=etag,
|
|
format=Path(nc_file_path).suffix.lstrip(".").lower(),
|
|
file_size_bytes=meta.size,
|
|
),
|
|
member_id,
|
|
)
|
|
log.info("Imported '%s' as version %s for song '%s'", nc_file_path, version.id, song.title)
|
|
|
|
imported += 1
|
|
read = SongRead.model_validate(song).model_copy(
|
|
update={"version_count": 1, "session_id": rehearsal_session_id}
|
|
)
|
|
yield {"type": "song", "song": read.model_dump(mode="json"), "is_new": is_new}
|
|
|
|
except Exception as exc:
|
|
log.error("Failed to import '%s': %s", nc_file_path, exc, exc_info=True)
|
|
skipped += 1
|
|
yield {"type": "skipped", "path": nc_file_path, "reason": f"import error: {exc}"}
|
|
|
|
yield {
|
|
"type": "done",
|
|
"stats": {"found": found, "imported": imported, "skipped": skipped},
|
|
}
|