examples-auto-run What it does Runs uv run examples/run_examples.py with: EXAMPLES_INTERACTIVE_MODE=auto (auto-input/auto-approve). Per-example logs under .tmp/examples-start-logs/. Main summary log path passed via --main-log (also under .tmp/examples-start-logs/). Generates a rerun list of failures at .tmp/examples-rerun.txt when --write-rerun is set. Provides start/stop/status/logs/tail/collect/rerun helpers via run.sh. Background option keeps the process running with a pidfile; stop cleans it up. Usage
Start (auto mode; interactive included by default)
.codex/skills/examples-auto-run/scripts/run.sh start [extra args to run_examples.py]
Examples:
.codex/skills/examples-auto-run/scripts/run.sh start --filter basic .codex/skills/examples-auto-run/scripts/run.sh start --include-server --include-audio
Check status
.codex/skills/examples-auto-run/scripts/run.sh status
Stop running job
.codex/skills/examples-auto-run/scripts/run.sh stop
List logs
.codex/skills/examples-auto-run/scripts/run.sh logs
Tail latest log (or specify one)
.codex/skills/examples-auto-run/scripts/run.sh tail .codex/skills/examples-auto-run/scripts/run.sh tail main_20260113-123000.log
Collect rerun list from a main log (defaults to latest main_*.log)
.codex/skills/examples-auto-run/scripts/run.sh collect
Rerun only failed entries from rerun file (auto mode)
.codex/skills/examples-auto-run/scripts/run.sh rerun
Defaults (overridable via env)
EXAMPLES_INTERACTIVE_MODE=auto
EXAMPLES_INCLUDE_INTERACTIVE=1
EXAMPLES_INCLUDE_SERVER=0
EXAMPLES_INCLUDE_AUDIO=0
EXAMPLES_INCLUDE_EXTERNAL=0
Auto-approvals in auto mode: APPLY_PATCH_AUTO_APPROVE=1, SHELL_AUTO_APPROVE=1, AUTO_APPROVE_MCP=1
Log locations
Main logs: .tmp/examples-start-logs/main_.log
Per-example logs (from run_examples.py): .tmp/examples-start-logs/
The runner does not perform any automated behavioral validation. After every foreground start or rerun, Codex must manually validate all exit-0 entries:
Read the example source (and comments) to infer intended flow, tools used, and expected key outputs. Open the matching per-example log under .tmp/examples-start-logs/. Confirm the intended actions/results occurred; flag omissions or divergences. Do this for all passed examples, not just a sample. Report immediately after the run with concise citations to the exact log lines that justify the validation.