Files
tapehoard/backend/tests/test_service_scanner.py
T
adamlamers 0bfd5affb4
Continuous Integration / backend-tests (push) Successful in 11m41s
Continuous Integration / frontend-check (push) Successful in 10m54s
tests & remove is_indexed flag from db
2026-04-28 23:55:39 -04:00

158 lines
4.7 KiB
Python

import hashlib
from datetime import datetime, timezone
from app.services.scanner import ScannerService, JobManager
from app.db import models
def test_job_manager_lifecycle(db_session):
"""Tests the full lifecycle of a job through JobManager."""
job = JobManager.create_job(db_session, "SCAN")
assert job.status == "PENDING"
JobManager.start_job(job.id)
# Refresh to see changes
db_session.expire_all()
job = db_session.get(models.Job, job.id)
assert job.status == "RUNNING"
assert job.started_at is not None
JobManager.update_job(job.id, 50.0, "Processing metadata")
db_session.expire_all()
job = db_session.get(models.Job, job.id)
assert job.progress == 50.0
assert job.current_task == "Processing metadata"
JobManager.complete_job(job.id)
db_session.expire_all()
job = db_session.get(models.Job, job.id)
assert job.status == "COMPLETED"
assert job.progress == 100.0
def test_job_manager_cancellation(db_session):
"""Tests job cancellation logic."""
job = JobManager.create_job(db_session, "HASH")
JobManager.cancel_job(job.id)
assert JobManager.is_cancelled(job.id) is True
db_session.expire_all()
job = db_session.get(models.Job, job.id)
assert job.status == "FAILED"
assert "Cancelled" in job.error_message
def test_compute_sha256(tmp_path):
"""Tests SHA-256 computation with actual file."""
scanner = ScannerService()
test_file = tmp_path / "test.bin"
content = b"tapehoard test content"
test_file.write_bytes(content)
expected_hash = hashlib.sha256(content).hexdigest()
actual_hash = scanner.compute_sha256(str(test_file))
assert actual_hash == expected_hash
def test_metadata_sync_batch(db_session):
"""Tests batch metadata synchronization logic."""
scanner = ScannerService()
timestamp = datetime.now(timezone.utc)
batch = [
{"path": "/data/f1.txt", "size": 100, "mtime": 1000, "ignored": False},
{"path": "/data/f2.txt", "size": 200, "mtime": 2000, "ignored": True},
]
scanner._sync_metadata_batch(db_session, batch, timestamp)
db_session.commit()
# Verify records created
f1 = (
db_session.query(models.FilesystemState)
.filter_by(file_path="/data/f1.txt")
.first()
)
assert f1.size == 100
assert f1.is_ignored is False
f2 = (
db_session.query(models.FilesystemState)
.filter_by(file_path="/data/f2.txt")
.first()
)
assert f2.is_ignored is True
def test_metadata_update_on_change(db_session):
"""Tests that existing metadata is updated if size/mtime change."""
scanner = ScannerService()
timestamp = datetime.now(timezone.utc)
# Initial state
f1 = models.FilesystemState(
file_path="/data/up.txt", size=50, mtime=1, sha256_hash="old"
)
db_session.add(f1)
db_session.commit()
# Update: size changed
batch = [{"path": "/data/up.txt", "size": 999, "mtime": 1, "ignored": False}]
scanner._sync_metadata_batch(db_session, batch, timestamp)
db_session.commit()
db_session.refresh(f1)
assert f1.size == 999
assert f1.sha256_hash is None # Should be reset for re-hashing
def test_scan_sources_mocked(db_session, mocker):
"""Tests the discovery scan with mocked filesystem."""
scanner = ScannerService()
# Mock settings
mocker.patch("app.api.system.get_source_roots", return_value=["/mock_source"])
mocker.patch("app.api.system.get_exclusion_spec", return_value=None)
# Mock os.walk and os.stat
mocker.patch("os.path.exists", return_value=True)
mocker.patch("os.walk", return_value=[("/mock_source", ["subdir"], ["file.txt"])])
mock_stat = mocker.MagicMock()
mock_stat.st_size = 500
mock_stat.st_mtime = 12345
mocker.patch("os.stat", return_value=mock_stat)
scanner.scan_sources(db_session)
# Verify file was discovered
record = (
db_session.query(models.FilesystemState)
.filter_by(file_path="/mock_source/file.txt")
.first()
)
assert record is not None
assert record.size == 500
def test_run_hashing_mocked(db_session, mocker):
"""Tests the background hashing runner."""
scanner = ScannerService()
# Setup unindexed file
f = models.FilesystemState(
file_path="/data/hash.me", size=10, mtime=1, is_ignored=False
)
db_session.add(f)
db_session.commit()
# Mock compute_sha256 to return a fixed hash
mocker.patch.object(ScannerService, "compute_sha256", return_value="mocked_hash")
# run_hashing runs in a loop until work is done.
# Since we aren't in 'is_running' state, it should process the 1 file and stop.
scanner.run_hashing()
db_session.refresh(f)
assert f.sha256_hash == "mocked_hash"