Skip to main content

XRTM Operator Runbook

This path covers the supported operating workflow for the top-level XRTM product shell once the released first event-forecasting loop is already working.

Page role

Released user workflow. This page stays on the published operator surface. For upcoming shortcuts or validation commands that are intentionally unreleased, use Next release track. For package-level runtime work, use Developer / integrator workflow.

It sharpens the official monitoring, history, and report workflow, while also covering the released benchmark smoke command that should precede local-LLM use.

Released xrtm==0.3.1 keeps this page on the commands that actually ship today. Corpus-validation flows remain on the next-release track.

Start from the default path

If you have not already done so, complete Getting started first. Operators should start with the same provider-free proof path as everyone else.

Official proof-point coverage

  1. Provider-free first success starts in Getting started.
  2. Benchmark smoke workflow appears below through xrtm perf run.
  3. Monitoring, history, and report workflow is the main operator loop on this page.
  4. Local-LLM advanced workflow stays optional and comes last.

Core operator loop

Create a repeatable profile

xrtm profile starter my-local --runs-dir runs
xrtm run profile my-local

Inspect artifacts and reports

xrtm runs list --runs-dir runs
xrtm runs show latest --runs-dir runs
xrtm artifacts inspect --latest --runs-dir runs
xrtm report html --latest --runs-dir runs
xrtm runs compare <run-id-a> <run-id-b> --runs-dir runs
xrtm runs export latest --runs-dir runs --output export.json
xrtm runs export latest --runs-dir runs --output export.csv --format csv

Browse the same runs in browser or terminal

xrtm web --runs-dir runs
xrtm tui --runs-dir runs

Use local monitoring and retention tools

xrtm monitor start --provider mock --limit 2 --runs-dir runs
xrtm monitor list --runs-dir runs
xrtm artifacts cleanup --runs-dir runs --keep 50

Monitor-only views read real monitor runs. Ordinary forecast runs stay in xrtm runs list, the WebUI runs table, and the TUI runs panel unless they were created with xrtm monitor start.

History and report review stay on the same proof-point path:

xrtm runs compare <run-id-a> <run-id-b> --runs-dir runs
xrtm runs export latest --runs-dir runs --output export.json
xrtm runs export latest --runs-dir runs --output export.csv --format csv
xrtm report html --latest --runs-dir runs

Treat that compare/export surface as an operator decision loop:

  • repeated mock runs are the stable default control
  • unchanged compare output means your baseline is behaving as expected
  • only claim improvement after a meaningful provider/model/runtime change lowers Brier/ECE without adding warnings/errors or unacceptable runtime/tokens cost
  • export the winning run when you want deeper analyst review

Benchmark smoke workflow

Use the built-in provider-free performance harness when you want a quick local regression signal without introducing provider noise:

xrtm perf run --scenario provider-free-smoke --iterations 3 --limit 1 --runs-dir runs-perf --output performance.json
xrtm web --runs-dir runs --smoke

xrtm perf run writes a structured performance report, while xrtm web --runs-dir runs --smoke verifies the WebUI routes without starting a long-lived server.

That benchmark is the honest released default baseline. Stronger visible improvement proof belongs to the advanced paths where you intentionally change model or runtime behavior.

Validation status

The newer corpus-validation commands are not part of the published xrtm==0.3.1 release. Until a later coordinated release ships them with compatible upstream packages, keep released operator workflows on xrtm perf run, explicit run inspection, compare, JSON/CSV export, and monitor commands.

Optional later: local-LLM mode

Use local-LLM mode only after the provider-free path above is already healthy.

export XRTM_LOCAL_LLM_BASE_URL=http://localhost:8080/v1
xrtm local-llm status
xrtm demo --provider local-llm --limit 1 --max-tokens 768 --runs-dir runs-local

This is the right place to validate a real local model, but it is intentionally not the first-run path.

Troubleshooting cues

  • xrtm install fails on Python 3.13: use Python 3.11 or 3.12.
  • xrtm local-llm status fails: treat it as a local endpoint/server issue first, then retry the local-LLM flow.
  • Need a run id: use xrtm runs list --runs-dir runs.
  • xrtm artifacts inspect fails on a directory: confirm it is a canonical XRTM run with run.json present.

What this runbook assumes today

  • local artifact-backed runs
  • profile-based repeatability via xrtm profile create
  • WebUI and TUI over the same run directories
  • compare, export, and inspect commands
  • monitor lifecycle commands
  • performance smoke checks