phase cv 0
This commit is contained in:
@@ -1,16 +1,20 @@
|
||||
"""
|
||||
API endpoints for checkpoint inspection, replay, and retry.
|
||||
API endpoints for checkpoint inspection, replay, retry, and GPU proxy.
|
||||
|
||||
GET /detect/checkpoints/{job_id} — list available checkpoints
|
||||
POST /detect/replay — replay from a stage with config overrides
|
||||
POST /detect/retry — queue async retry with different provider
|
||||
POST /detect/replay-stage — replay single stage (fast path)
|
||||
POST /detect/gpu/detect_edges — proxy to GPU inference server
|
||||
POST /detect/gpu/detect_edges/debug — proxy with debug overlays
|
||||
"""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import logging
|
||||
import os
|
||||
|
||||
from fastapi import APIRouter, HTTPException
|
||||
from fastapi import APIRouter, HTTPException, Request, Response
|
||||
from pydantic import BaseModel
|
||||
|
||||
logger = logging.getLogger(__name__)
|
||||
@@ -22,6 +26,18 @@ router = APIRouter(prefix="/detect", tags=["detect"])
|
||||
|
||||
class CheckpointInfo(BaseModel):
|
||||
stage: str
|
||||
is_scenario: bool = False
|
||||
scenario_label: str = ""
|
||||
|
||||
|
||||
class ScenarioInfo(BaseModel):
|
||||
job_id: str
|
||||
stage: str
|
||||
scenario_label: str
|
||||
profile_name: str
|
||||
video_path: str
|
||||
frame_count: int = 0
|
||||
created_at: str = ""
|
||||
|
||||
|
||||
class ReplayRequest(BaseModel):
|
||||
@@ -51,6 +67,39 @@ class RetryResponse(BaseModel):
|
||||
job_id: str
|
||||
|
||||
|
||||
class ReplaySingleStageRequest(BaseModel):
|
||||
job_id: str
|
||||
stage: str
|
||||
frame_refs: list[int] | None = None
|
||||
config_overrides: dict | None = None
|
||||
debug: bool = False
|
||||
|
||||
|
||||
class ReplaySingleStageBox(BaseModel):
|
||||
x: int
|
||||
y: int
|
||||
w: int
|
||||
h: int
|
||||
confidence: float
|
||||
label: str
|
||||
|
||||
|
||||
class FrameDebugOverlays(BaseModel):
|
||||
edge_overlay_b64: str = ""
|
||||
lines_overlay_b64: str = ""
|
||||
horizontal_count: int = 0
|
||||
pair_count: int = 0
|
||||
|
||||
|
||||
class ReplaySingleStageResponse(BaseModel):
|
||||
status: str
|
||||
stage: str
|
||||
frame_count: int = 0
|
||||
region_count: int = 0
|
||||
regions_by_frame: dict[str, list[ReplaySingleStageBox]] = {}
|
||||
debug: dict[str, FrameDebugOverlays] = {} # keyed by frame seq
|
||||
|
||||
|
||||
# --- Endpoints ---
|
||||
|
||||
@router.get("/checkpoints/{job_id}")
|
||||
@@ -67,6 +116,28 @@ def list_checkpoints(job_id: str) -> list[CheckpointInfo]:
|
||||
return result
|
||||
|
||||
|
||||
@router.get("/scenarios", response_model=list[ScenarioInfo])
|
||||
def list_scenarios_endpoint():
|
||||
"""List all available scenarios (bookmarked checkpoints)."""
|
||||
from core.db.detect import list_scenarios
|
||||
|
||||
scenarios = list_scenarios()
|
||||
result = []
|
||||
for s in scenarios:
|
||||
manifest = s.frames_manifest or {}
|
||||
info = ScenarioInfo(
|
||||
job_id=str(s.job_id),
|
||||
stage=s.stage,
|
||||
scenario_label=s.scenario_label,
|
||||
profile_name=s.profile_name,
|
||||
video_path=s.video_path,
|
||||
frame_count=len(manifest),
|
||||
created_at=str(s.created_at) if s.created_at else "",
|
||||
)
|
||||
result.append(info)
|
||||
return result
|
||||
|
||||
|
||||
@router.post("/replay", response_model=ReplayResponse)
|
||||
def replay(req: ReplayRequest):
|
||||
"""Replay pipeline from a specific stage with optional config overrides."""
|
||||
@@ -119,3 +190,103 @@ def retry(req: RetryRequest):
|
||||
job_id=req.job_id,
|
||||
)
|
||||
return response
|
||||
|
||||
|
||||
@router.post("/replay-stage", response_model=ReplaySingleStageResponse)
|
||||
def replay_single_stage(req: ReplaySingleStageRequest):
|
||||
"""Replay a single stage on specific frames — fast path for interactive tuning."""
|
||||
from detect.checkpoint.replay import replay_single_stage as _replay
|
||||
|
||||
try:
|
||||
result = _replay(
|
||||
job_id=req.job_id,
|
||||
stage=req.stage,
|
||||
frame_refs=req.frame_refs,
|
||||
config_overrides=req.config_overrides,
|
||||
debug=req.debug,
|
||||
)
|
||||
except ValueError as e:
|
||||
raise HTTPException(status_code=400, detail=str(e))
|
||||
except Exception as e:
|
||||
raise HTTPException(status_code=500, detail=f"Single-stage replay failed: {e}")
|
||||
|
||||
# Convert result to response format
|
||||
regions_by_frame = result.get("edge_regions_by_frame", {})
|
||||
total_regions = 0
|
||||
serialized = {}
|
||||
for seq, boxes in regions_by_frame.items():
|
||||
box_list = []
|
||||
for b in boxes:
|
||||
box = ReplaySingleStageBox(
|
||||
x=b.x, y=b.y, w=b.w, h=b.h,
|
||||
confidence=b.confidence, label=b.label,
|
||||
)
|
||||
box_list.append(box)
|
||||
serialized[str(seq)] = box_list
|
||||
total_regions += len(box_list)
|
||||
|
||||
# Serialize debug overlays if present
|
||||
debug_out = {}
|
||||
raw_debug = result.get("debug", {})
|
||||
for seq, d in raw_debug.items():
|
||||
debug_out[str(seq)] = FrameDebugOverlays(
|
||||
edge_overlay_b64=d.get("edge_overlay_b64", ""),
|
||||
lines_overlay_b64=d.get("lines_overlay_b64", ""),
|
||||
horizontal_count=d.get("horizontal_count", 0),
|
||||
pair_count=d.get("pair_count", 0),
|
||||
)
|
||||
|
||||
return ReplaySingleStageResponse(
|
||||
status="completed",
|
||||
stage=req.stage,
|
||||
frame_count=len(regions_by_frame),
|
||||
region_count=total_regions,
|
||||
regions_by_frame=serialized,
|
||||
debug=debug_out,
|
||||
)
|
||||
|
||||
|
||||
# --- GPU proxy — thin passthrough to inference server for interactive editor ---
|
||||
|
||||
|
||||
def _gpu_url() -> str:
|
||||
url = os.environ.get("INFERENCE_URL", "http://localhost:8000")
|
||||
return url.rstrip("/")
|
||||
|
||||
|
||||
@router.post("/gpu/detect_edges")
|
||||
async def gpu_detect_edges(request: Request):
|
||||
"""Proxy to GPU inference server — browser can't reach it directly."""
|
||||
import httpx
|
||||
|
||||
body = await request.body()
|
||||
try:
|
||||
async with httpx.AsyncClient(timeout=30.0) as client:
|
||||
resp = await client.post(
|
||||
f"{_gpu_url()}/detect_edges",
|
||||
content=body,
|
||||
headers={"Content-Type": "application/json"},
|
||||
)
|
||||
return Response(content=resp.content, status_code=resp.status_code,
|
||||
media_type="application/json")
|
||||
except Exception as e:
|
||||
raise HTTPException(status_code=502, detail=f"GPU server unreachable: {e}")
|
||||
|
||||
|
||||
@router.post("/gpu/detect_edges/debug")
|
||||
async def gpu_detect_edges_debug(request: Request):
|
||||
"""Proxy to GPU inference server debug endpoint."""
|
||||
import httpx
|
||||
|
||||
body = await request.body()
|
||||
try:
|
||||
async with httpx.AsyncClient(timeout=30.0) as client:
|
||||
resp = await client.post(
|
||||
f"{_gpu_url()}/detect_edges/debug",
|
||||
content=body,
|
||||
headers={"Content-Type": "application/json"},
|
||||
)
|
||||
return Response(content=resp.content, status_code=resp.status_code,
|
||||
media_type="application/json")
|
||||
except Exception as e:
|
||||
raise HTTPException(status_code=502, detail=f"GPU server unreachable: {e}")
|
||||
|
||||
Reference in New Issue
Block a user