Reliability Baseline¶
This page defines the minimum reliability signals LoopForge tracks during onboarding.
Why this exists¶
The goal is simple:
- reduce time-to-first-success
- make onboarding failures diagnosable
- give both users and maintainers the same evidence trail
Core onboarding artifacts¶
A successful or failed onboarding flow can produce three useful artifact types:
~/.loopforge/onboard-metrics.json~/.loopforge/onboard-events.jsonl<workspace>/.loopforge/onboard-report.jsonand.md
Use them together:
- metrics show trends
- events show raw attempts
- the workspace report shows the last run's setup status, first-task status, and next actions
Core metrics¶
LoopForge tracks:
attempted_first_taskfirst_task_successfirst_task_failedfailure_by_category
Typical categories:
model_unavailableprovider_unreachabletool_runtime_errorsandbox_restrictionunknown
Daily report script¶
LoopForge includes a reporting helper:
scripts/onboard_metrics_report.py
Run it from repository root:
It generates:
.tmp/onboard-report/onboard-report.json.tmp/onboard-report/onboard-report.md
The Markdown report now includes:
- metrics snapshot
- recent failure categories
- recommended fixes for top failure types
- daily trend table
Operational loop¶
- Run
loopforge onboard - Read
.loopforge/onboard-report.mdin the workspace - Run
loopforge doctorif needed - Aggregate trends with
scripts/onboard_metrics_report.py - Fix the top repeated failure category first