fase 7 afgeronf
This commit is contained in:
@@ -170,3 +170,26 @@ def rename_execute(
|
|||||||
return service.execute_rename(normalized_session_id, confirm=confirm)
|
return service.execute_rename(normalized_session_id, confirm=confirm)
|
||||||
except ValueError as exc:
|
except ValueError as exc:
|
||||||
raise HTTPException(status_code=400, detail=str(exc))
|
raise HTTPException(status_code=400, detail=str(exc))
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/rename-log")
|
||||||
|
def get_rename_log(
|
||||||
|
session_id: str = Query("default", min_length=1),
|
||||||
|
limit: int = Query(20, ge=1, le=200),
|
||||||
|
):
|
||||||
|
service = SessionService()
|
||||||
|
normalized_session_id = _normalize_session_id(session_id)
|
||||||
|
items = service.list_rename_runs(normalized_session_id, limit=limit)
|
||||||
|
return {
|
||||||
|
"session_id": normalized_session_id,
|
||||||
|
"items": items,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@router.get("/rename-log/{run_id}")
|
||||||
|
def get_rename_log_run(run_id: int):
|
||||||
|
service = SessionService()
|
||||||
|
run = service.get_rename_run(run_id)
|
||||||
|
if run is None:
|
||||||
|
raise HTTPException(status_code=404, detail="rename run not found")
|
||||||
|
return run
|
||||||
|
|||||||
@@ -1,6 +1,8 @@
|
|||||||
import json
|
import json
|
||||||
import os
|
import os
|
||||||
import sqlite3
|
import sqlite3
|
||||||
|
import time
|
||||||
|
from datetime import datetime, timezone
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from app.config import APP_DATA_DIR
|
from app.config import APP_DATA_DIR
|
||||||
@@ -63,6 +65,50 @@ class SessionService:
|
|||||||
ON selected_files(session_id)
|
ON selected_files(session_id)
|
||||||
"""
|
"""
|
||||||
)
|
)
|
||||||
|
conn.execute(
|
||||||
|
"""
|
||||||
|
CREATE TABLE IF NOT EXISTS rename_runs (
|
||||||
|
run_id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
session_id TEXT NOT NULL,
|
||||||
|
created_at TEXT NOT NULL,
|
||||||
|
confirm INTEGER NOT NULL,
|
||||||
|
executed INTEGER NOT NULL,
|
||||||
|
preflight_ok INTEGER NOT NULL,
|
||||||
|
episodes_count INTEGER NOT NULL,
|
||||||
|
files_count INTEGER NOT NULL,
|
||||||
|
duration_ms INTEGER NOT NULL
|
||||||
|
)
|
||||||
|
"""
|
||||||
|
)
|
||||||
|
conn.execute(
|
||||||
|
"""
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_rename_runs_session
|
||||||
|
ON rename_runs(session_id)
|
||||||
|
"""
|
||||||
|
)
|
||||||
|
conn.execute(
|
||||||
|
"""
|
||||||
|
CREATE TABLE IF NOT EXISTS rename_run_items (
|
||||||
|
id INTEGER PRIMARY KEY AUTOINCREMENT,
|
||||||
|
run_id INTEGER NOT NULL,
|
||||||
|
item_index INTEGER NOT NULL,
|
||||||
|
episode_selection_id INTEGER,
|
||||||
|
file_selection_id INTEGER,
|
||||||
|
source_path TEXT NOT NULL,
|
||||||
|
destination_path TEXT NOT NULL,
|
||||||
|
proposed_filename TEXT NOT NULL,
|
||||||
|
status TEXT NOT NULL,
|
||||||
|
errors_json TEXT NOT NULL,
|
||||||
|
FOREIGN KEY(run_id) REFERENCES rename_runs(run_id)
|
||||||
|
)
|
||||||
|
"""
|
||||||
|
)
|
||||||
|
conn.execute(
|
||||||
|
"""
|
||||||
|
CREATE INDEX IF NOT EXISTS idx_rename_run_items_run
|
||||||
|
ON rename_run_items(run_id, item_index)
|
||||||
|
"""
|
||||||
|
)
|
||||||
|
|
||||||
def list_selected_episodes(self, session_id: str) -> list[dict]:
|
def list_selected_episodes(self, session_id: str) -> list[dict]:
|
||||||
with self._connect() as conn:
|
with self._connect() as conn:
|
||||||
@@ -407,6 +453,7 @@ class SessionService:
|
|||||||
if not confirm:
|
if not confirm:
|
||||||
raise ValueError("confirm=true is required to execute rename")
|
raise ValueError("confirm=true is required to execute rename")
|
||||||
|
|
||||||
|
started_at = time.perf_counter()
|
||||||
preview = self.build_filename_preview(session_id)
|
preview = self.build_filename_preview(session_id)
|
||||||
allowed_roots = self._allowed_media_roots()
|
allowed_roots = self._allowed_media_roots()
|
||||||
preflight_items = []
|
preflight_items = []
|
||||||
@@ -444,7 +491,7 @@ class SessionService:
|
|||||||
)
|
)
|
||||||
|
|
||||||
if preflight_errors > 0:
|
if preflight_errors > 0:
|
||||||
return {
|
result = {
|
||||||
"session_id": session_id,
|
"session_id": session_id,
|
||||||
"confirm": confirm,
|
"confirm": confirm,
|
||||||
"executed": False,
|
"executed": False,
|
||||||
@@ -452,6 +499,12 @@ class SessionService:
|
|||||||
"counts": preview["counts"],
|
"counts": preview["counts"],
|
||||||
"items": preflight_items,
|
"items": preflight_items,
|
||||||
}
|
}
|
||||||
|
self._log_rename_run(
|
||||||
|
session_id=session_id,
|
||||||
|
result=result,
|
||||||
|
duration_ms=int((time.perf_counter() - started_at) * 1000),
|
||||||
|
)
|
||||||
|
return result
|
||||||
|
|
||||||
results = []
|
results = []
|
||||||
for item in preflight_items:
|
for item in preflight_items:
|
||||||
@@ -466,7 +519,7 @@ class SessionService:
|
|||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
return {
|
result = {
|
||||||
"session_id": session_id,
|
"session_id": session_id,
|
||||||
"confirm": confirm,
|
"confirm": confirm,
|
||||||
"executed": True,
|
"executed": True,
|
||||||
@@ -474,6 +527,204 @@ class SessionService:
|
|||||||
"counts": preview["counts"],
|
"counts": preview["counts"],
|
||||||
"items": results,
|
"items": results,
|
||||||
}
|
}
|
||||||
|
self._log_rename_run(
|
||||||
|
session_id=session_id,
|
||||||
|
result=result,
|
||||||
|
duration_ms=int((time.perf_counter() - started_at) * 1000),
|
||||||
|
)
|
||||||
|
return result
|
||||||
|
|
||||||
|
def _log_rename_run(self, session_id: str, result: dict, duration_ms: int) -> None:
|
||||||
|
created_at = datetime.now(timezone.utc).isoformat()
|
||||||
|
counts = result.get("counts", {})
|
||||||
|
|
||||||
|
with self._connect() as conn:
|
||||||
|
cursor = conn.execute(
|
||||||
|
"""
|
||||||
|
INSERT INTO rename_runs (
|
||||||
|
session_id,
|
||||||
|
created_at,
|
||||||
|
confirm,
|
||||||
|
executed,
|
||||||
|
preflight_ok,
|
||||||
|
episodes_count,
|
||||||
|
files_count,
|
||||||
|
duration_ms
|
||||||
|
)
|
||||||
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?)
|
||||||
|
""",
|
||||||
|
(
|
||||||
|
session_id,
|
||||||
|
created_at,
|
||||||
|
1 if result.get("confirm") else 0,
|
||||||
|
1 if result.get("executed") else 0,
|
||||||
|
1 if result.get("preflight_ok") else 0,
|
||||||
|
int(counts.get("episodes") or 0),
|
||||||
|
int(counts.get("files") or 0),
|
||||||
|
int(duration_ms),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
run_id = int(cursor.lastrowid)
|
||||||
|
|
||||||
|
for item in result.get("items", []):
|
||||||
|
conn.execute(
|
||||||
|
"""
|
||||||
|
INSERT INTO rename_run_items (
|
||||||
|
run_id,
|
||||||
|
item_index,
|
||||||
|
episode_selection_id,
|
||||||
|
file_selection_id,
|
||||||
|
source_path,
|
||||||
|
destination_path,
|
||||||
|
proposed_filename,
|
||||||
|
status,
|
||||||
|
errors_json
|
||||||
|
)
|
||||||
|
VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)
|
||||||
|
""",
|
||||||
|
(
|
||||||
|
run_id,
|
||||||
|
int(item.get("index") or 0),
|
||||||
|
item.get("episode_selection_id"),
|
||||||
|
item.get("file_selection_id"),
|
||||||
|
str(item.get("source_path") or ""),
|
||||||
|
str(item.get("destination_path") or ""),
|
||||||
|
str(item.get("proposed_filename") or ""),
|
||||||
|
str(item.get("status") or ""),
|
||||||
|
json.dumps(item.get("errors") or [], ensure_ascii=True),
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
def list_rename_runs(self, session_id: str, limit: int = 20) -> list[dict]:
|
||||||
|
with self._connect() as conn:
|
||||||
|
rows = conn.execute(
|
||||||
|
"""
|
||||||
|
SELECT
|
||||||
|
run_id,
|
||||||
|
session_id,
|
||||||
|
created_at,
|
||||||
|
confirm,
|
||||||
|
executed,
|
||||||
|
preflight_ok,
|
||||||
|
episodes_count,
|
||||||
|
files_count,
|
||||||
|
duration_ms
|
||||||
|
FROM rename_runs
|
||||||
|
WHERE session_id = ?
|
||||||
|
ORDER BY run_id DESC
|
||||||
|
LIMIT ?
|
||||||
|
""",
|
||||||
|
(session_id, limit),
|
||||||
|
).fetchall()
|
||||||
|
|
||||||
|
item_counts = conn.execute(
|
||||||
|
"""
|
||||||
|
SELECT run_id, COUNT(*) AS item_count
|
||||||
|
FROM rename_run_items
|
||||||
|
WHERE run_id IN (
|
||||||
|
SELECT run_id
|
||||||
|
FROM rename_runs
|
||||||
|
WHERE session_id = ?
|
||||||
|
ORDER BY run_id DESC
|
||||||
|
LIMIT ?
|
||||||
|
)
|
||||||
|
GROUP BY run_id
|
||||||
|
""",
|
||||||
|
(session_id, limit),
|
||||||
|
).fetchall()
|
||||||
|
|
||||||
|
counts_by_run = {int(row["run_id"]): int(row["item_count"]) for row in item_counts}
|
||||||
|
runs = []
|
||||||
|
for row in rows:
|
||||||
|
run_id = int(row["run_id"])
|
||||||
|
runs.append(
|
||||||
|
{
|
||||||
|
"run_id": run_id,
|
||||||
|
"session_id": row["session_id"],
|
||||||
|
"created_at": row["created_at"],
|
||||||
|
"confirm": bool(row["confirm"]),
|
||||||
|
"executed": bool(row["executed"]),
|
||||||
|
"preflight_ok": bool(row["preflight_ok"]),
|
||||||
|
"counts": {
|
||||||
|
"episodes": int(row["episodes_count"]),
|
||||||
|
"files": int(row["files_count"]),
|
||||||
|
},
|
||||||
|
"duration_ms": int(row["duration_ms"]),
|
||||||
|
"items_count": counts_by_run.get(run_id, 0),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
return runs
|
||||||
|
|
||||||
|
def get_rename_run(self, run_id: int) -> dict | None:
|
||||||
|
with self._connect() as conn:
|
||||||
|
run = conn.execute(
|
||||||
|
"""
|
||||||
|
SELECT
|
||||||
|
run_id,
|
||||||
|
session_id,
|
||||||
|
created_at,
|
||||||
|
confirm,
|
||||||
|
executed,
|
||||||
|
preflight_ok,
|
||||||
|
episodes_count,
|
||||||
|
files_count,
|
||||||
|
duration_ms
|
||||||
|
FROM rename_runs
|
||||||
|
WHERE run_id = ?
|
||||||
|
""",
|
||||||
|
(run_id,),
|
||||||
|
).fetchone()
|
||||||
|
|
||||||
|
if run is None:
|
||||||
|
return None
|
||||||
|
|
||||||
|
item_rows = conn.execute(
|
||||||
|
"""
|
||||||
|
SELECT
|
||||||
|
item_index,
|
||||||
|
episode_selection_id,
|
||||||
|
file_selection_id,
|
||||||
|
source_path,
|
||||||
|
destination_path,
|
||||||
|
proposed_filename,
|
||||||
|
status,
|
||||||
|
errors_json
|
||||||
|
FROM rename_run_items
|
||||||
|
WHERE run_id = ?
|
||||||
|
ORDER BY item_index ASC
|
||||||
|
""",
|
||||||
|
(run_id,),
|
||||||
|
).fetchall()
|
||||||
|
|
||||||
|
items = []
|
||||||
|
for row in item_rows:
|
||||||
|
items.append(
|
||||||
|
{
|
||||||
|
"index": int(row["item_index"]),
|
||||||
|
"episode_selection_id": row["episode_selection_id"],
|
||||||
|
"file_selection_id": row["file_selection_id"],
|
||||||
|
"source_path": row["source_path"],
|
||||||
|
"destination_path": row["destination_path"],
|
||||||
|
"proposed_filename": row["proposed_filename"],
|
||||||
|
"status": row["status"],
|
||||||
|
"errors": json.loads(row["errors_json"]),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
return {
|
||||||
|
"run_id": int(run["run_id"]),
|
||||||
|
"session_id": run["session_id"],
|
||||||
|
"created_at": run["created_at"],
|
||||||
|
"confirm": bool(run["confirm"]),
|
||||||
|
"executed": bool(run["executed"]),
|
||||||
|
"preflight_ok": bool(run["preflight_ok"]),
|
||||||
|
"counts": {
|
||||||
|
"episodes": int(run["episodes_count"]),
|
||||||
|
"files": int(run["files_count"]),
|
||||||
|
},
|
||||||
|
"duration_ms": int(run["duration_ms"]),
|
||||||
|
"items": items,
|
||||||
|
}
|
||||||
|
|
||||||
def _allowed_media_roots(self) -> list[Path]:
|
def _allowed_media_roots(self) -> list[Path]:
|
||||||
raw = os.getenv("ALLOWED_MEDIA_ROOTS", "").strip()
|
raw = os.getenv("ALLOWED_MEDIA_ROOTS", "").strip()
|
||||||
|
|||||||
Binary file not shown.
Executable
+211
@@ -0,0 +1,211 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
if [ -z "${BASE_URL:-}" ]; then
|
||||||
|
if curl --silent --fail http://127.0.0.1:8085/api/health >/dev/null 2>&1; then
|
||||||
|
BASE_URL="http://127.0.0.1:8085"
|
||||||
|
elif curl --silent --fail http://host.containers.internal:8085/api/health >/dev/null 2>&1; then
|
||||||
|
BASE_URL="http://host.containers.internal:8085"
|
||||||
|
else
|
||||||
|
echo "ERROR: could not determine BASE_URL. Tried 127.0.0.1 and host.containers.internal." >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -z "${TEST_MEDIA_ROOT:-}" ]; then
|
||||||
|
for candidate in \
|
||||||
|
"/Volumes/8TB/Shared_Folders/TV_Shows" \
|
||||||
|
"/Volumes/8TB_RAID1/Shared_Folders/Library/TV_Shows"
|
||||||
|
do
|
||||||
|
if [ -d "$candidate" ] && [ -w "$candidate" ]; then
|
||||||
|
TEST_MEDIA_ROOT="$candidate"
|
||||||
|
break
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -z "${TEST_MEDIA_ROOT:-}" ]; then
|
||||||
|
echo "ERROR: no writable allowed media root found. Set TEST_MEDIA_ROOT." >&2
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
|
||||||
|
TEST_DIR="${TEST_MEDIA_ROOT}/_rename_mvp_phase7_$(date +%s)_$$"
|
||||||
|
mkdir -p "${TEST_DIR}"
|
||||||
|
|
||||||
|
TMP_DIR="$(mktemp -d)"
|
||||||
|
trap 'rm -rf "$TMP_DIR"' EXIT
|
||||||
|
|
||||||
|
clear_session() {
|
||||||
|
local sid="$1"
|
||||||
|
curl --fail --silent --show-error -X DELETE \
|
||||||
|
"${BASE_URL}/api/session/selected-episodes?session_id=${sid}" \
|
||||||
|
>/dev/null
|
||||||
|
curl --fail --silent --show-error -X DELETE \
|
||||||
|
"${BASE_URL}/api/session/selected-files?session_id=${sid}" \
|
||||||
|
>/dev/null
|
||||||
|
}
|
||||||
|
|
||||||
|
add_payloads() {
|
||||||
|
local sid="$1"
|
||||||
|
local file1="$2"
|
||||||
|
local file2="$3"
|
||||||
|
|
||||||
|
cat > "${TMP_DIR}/episodes_${sid}.json" <<'JSON'
|
||||||
|
{
|
||||||
|
"items": [
|
||||||
|
{
|
||||||
|
"id": 1,
|
||||||
|
"series": "Elsbeth",
|
||||||
|
"year": "2024",
|
||||||
|
"season_number": 1,
|
||||||
|
"episode_number": 1,
|
||||||
|
"title": "Pilot"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": 2,
|
||||||
|
"series": "Elsbeth",
|
||||||
|
"year": "2024",
|
||||||
|
"season_number": 1,
|
||||||
|
"episode_number": 2,
|
||||||
|
"title": "Second Episode"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
JSON
|
||||||
|
|
||||||
|
cat > "${TMP_DIR}/files_${sid}.json" <<JSON
|
||||||
|
{
|
||||||
|
"items": [
|
||||||
|
{
|
||||||
|
"path": "${file1}",
|
||||||
|
"name": "$(basename "${file1}")"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"path": "${file2}",
|
||||||
|
"name": "$(basename "${file2}")"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
JSON
|
||||||
|
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
-X POST "${BASE_URL}/api/session/selected-episodes?session_id=${sid}" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
--data @"${TMP_DIR}/episodes_${sid}.json" \
|
||||||
|
>/dev/null
|
||||||
|
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
-X POST "${BASE_URL}/api/session/selected-files?session_id=${sid}" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
--data @"${TMP_DIR}/files_${sid}.json" \
|
||||||
|
>/dev/null
|
||||||
|
}
|
||||||
|
|
||||||
|
echo "== Feature test 1: successful rename run is logged with duration =="
|
||||||
|
SESSION_OK="rename-log-ok-$(date +%s)-$$"
|
||||||
|
SRC1="${TEST_DIR}/ok_src1.mkv"
|
||||||
|
SRC2="${TEST_DIR}/ok_src2.mp4"
|
||||||
|
printf "a" > "${SRC1}"
|
||||||
|
printf "b" > "${SRC2}"
|
||||||
|
|
||||||
|
clear_session "${SESSION_OK}"
|
||||||
|
add_payloads "${SESSION_OK}" "${SRC1}" "${SRC2}"
|
||||||
|
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
-X POST "${BASE_URL}/api/session/rename-execute?session_id=${SESSION_OK}&confirm=true" \
|
||||||
|
-o "${TMP_DIR}/rename_ok.json"
|
||||||
|
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
"${BASE_URL}/api/session/rename-log?session_id=${SESSION_OK}" \
|
||||||
|
-o "${TMP_DIR}/log_ok.json"
|
||||||
|
|
||||||
|
cat "${TMP_DIR}/log_ok.json"
|
||||||
|
|
||||||
|
python3 - "${TMP_DIR}/log_ok.json" > "${TMP_DIR}/run_id_ok.txt" <<'PY'
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
data = json.loads(Path(sys.argv[1]).read_text(encoding="utf-8"))
|
||||||
|
items = data.get("items", [])
|
||||||
|
assert len(items) >= 1, "expected at least one logged run"
|
||||||
|
run = items[0]
|
||||||
|
assert run.get("executed") is True, "expected executed=true"
|
||||||
|
assert run.get("preflight_ok") is True, "expected preflight_ok=true"
|
||||||
|
assert isinstance(run.get("duration_ms"), int), "duration_ms must be int"
|
||||||
|
assert run["duration_ms"] >= 0, "duration_ms must be >= 0"
|
||||||
|
assert run.get("items_count") == 2, "expected 2 item logs"
|
||||||
|
print(run["run_id"])
|
||||||
|
PY
|
||||||
|
|
||||||
|
RUN_ID_OK="$(cat "${TMP_DIR}/run_id_ok.txt")"
|
||||||
|
|
||||||
|
echo
|
||||||
|
echo "== Feature test 2: preflight-failed run is also logged =="
|
||||||
|
SESSION_FAIL="rename-log-fail-$(date +%s)-$$"
|
||||||
|
SRC3="${TEST_DIR}/fail_src1.mkv"
|
||||||
|
SRC4="${TEST_DIR}/fail_src2.mp4"
|
||||||
|
printf "c" > "${SRC3}"
|
||||||
|
printf "d" > "${SRC4}"
|
||||||
|
|
||||||
|
# Force destination conflict for first mapping.
|
||||||
|
CONFLICT1="${TEST_DIR}/Elsbeth (2024) - S01E01 - Pilot.mkv"
|
||||||
|
printf "existing" > "${CONFLICT1}"
|
||||||
|
|
||||||
|
clear_session "${SESSION_FAIL}"
|
||||||
|
add_payloads "${SESSION_FAIL}" "${SRC3}" "${SRC4}"
|
||||||
|
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
-X POST "${BASE_URL}/api/session/rename-execute?session_id=${SESSION_FAIL}&confirm=true" \
|
||||||
|
-o "${TMP_DIR}/rename_fail.json"
|
||||||
|
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
"${BASE_URL}/api/session/rename-log?session_id=${SESSION_FAIL}" \
|
||||||
|
-o "${TMP_DIR}/log_fail.json"
|
||||||
|
|
||||||
|
cat "${TMP_DIR}/log_fail.json"
|
||||||
|
|
||||||
|
python3 - "${TMP_DIR}/log_fail.json" > "${TMP_DIR}/run_id_fail.txt" <<'PY'
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
data = json.loads(Path(sys.argv[1]).read_text(encoding="utf-8"))
|
||||||
|
items = data.get("items", [])
|
||||||
|
assert len(items) >= 1, "expected at least one logged run"
|
||||||
|
run = items[0]
|
||||||
|
assert run.get("executed") is False, "expected executed=false"
|
||||||
|
assert run.get("preflight_ok") is False, "expected preflight_ok=false"
|
||||||
|
assert isinstance(run.get("duration_ms"), int), "duration_ms must be int"
|
||||||
|
assert run["duration_ms"] >= 0, "duration_ms must be >= 0"
|
||||||
|
print(run["run_id"])
|
||||||
|
PY
|
||||||
|
|
||||||
|
RUN_ID_FAIL="$(cat "${TMP_DIR}/run_id_fail.txt")"
|
||||||
|
|
||||||
|
echo
|
||||||
|
echo "== Feature test 3: run detail endpoint returns item statuses and errors =="
|
||||||
|
curl --fail --silent --show-error \
|
||||||
|
"${BASE_URL}/api/session/rename-log/${RUN_ID_FAIL}" \
|
||||||
|
-o "${TMP_DIR}/run_detail_fail.json"
|
||||||
|
|
||||||
|
cat "${TMP_DIR}/run_detail_fail.json"
|
||||||
|
|
||||||
|
python3 - "${TMP_DIR}/run_detail_fail.json" <<'PY'
|
||||||
|
import json
|
||||||
|
import sys
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
data = json.loads(Path(sys.argv[1]).read_text(encoding="utf-8"))
|
||||||
|
assert data.get("executed") is False, "detail should show executed=false"
|
||||||
|
assert data.get("preflight_ok") is False, "detail should show preflight_ok=false"
|
||||||
|
assert isinstance(data.get("duration_ms"), int), "detail duration_ms must be int"
|
||||||
|
items = data.get("items", [])
|
||||||
|
assert len(items) == 2, "expected 2 item details"
|
||||||
|
assert any("destination file already exists" in " ".join(i.get("errors", [])) for i in items), \
|
||||||
|
"expected destination conflict error in item details"
|
||||||
|
print("rename-log detail validation passed")
|
||||||
|
PY
|
||||||
|
|
||||||
|
echo
|
||||||
|
echo "All rename log feature tests passed."
|
||||||
Reference in New Issue
Block a user