User Story
As Maya, in order to pick which LLM drives Carlos's conversational form-filling experience and see how different models perform on adaptive interviewing, I want filling variants to be selectable from Settings → Variants, backed by persona-driven evaluation.
Preconditions
Acceptance Criteria
Success Metrics
- Observable tradeoff across variants (e.g., Opus higher completeness, Haiku faster/cheaper)
- Persona-driven eval runs reproducibly from
bun run cli evaluate run <variant>
Notes
- Port evaluator shape from
llm-class-2026-winter-cohort/notes/assignment-10/ — persona-driven simulator + completeness scorer
- Filling registry stub exists at
src/services/forms/filling/registry.ts (empty); extend it
Definition of Done
User Story
As Maya, in order to pick which LLM drives Carlos's conversational form-filling experience and see how different models perform on adaptive interviewing, I want filling variants to be selectable from Settings → Variants, backed by persona-driven evaluation.
Preconditions
Acceptance Criteria
filling-interviewscores a variant against scripted personas (completeness of collected fields + conditional routing accuracy), mirroring the assignment-10 TextGrad evaluator patternfilling/haiku,filling/sonnet,filling/opus<VariantBadge task="filling" ...>rendered on conversation UIcatalog/experiments/filling-model-comparison/with_suite.md+ one markdown per variant, each with metrics + findingscatalog/experiments/_roadmap.mdupdated with shipped status and one-line findingSuccess Metrics
bun run cli evaluate run <variant>Notes
llm-class-2026-winter-cohort/notes/assignment-10/— persona-driven simulator + completeness scorersrc/services/forms/filling/registry.ts(empty); extend itDefinition of Done