Commit Graph

2 Commits

Author SHA1 Message Date
b309e7fd49 feat(eval-loop): Day 4 — LLM-assisted extractor path (additive, flagged)
Day 2 baseline showed 0% recall for the rule-based extractor across
5 distinct miss classes. Day 4 decision gate: prototype an
LLM-assisted mode behind a flag. Option A ratified by Antoine.

New module src/atocore/memory/extractor_llm.py:

- extract_candidates_llm(interaction) returns the same MemoryCandidate
  dataclass the rule extractor produces, so both paths flow through
  the existing triage / candidate pipeline unchanged.
- extract_candidates_llm_verbose() also returns the raw model output
  and any error string, for eval and debugging.
- Uses Claude Haiku 4.5 by default; model overridable via
  ATOCORE_LLM_EXTRACTOR_MODEL env. Timeout via
  ATOCORE_LLM_EXTRACTOR_TIMEOUT_S (default 20s).
- Silent-failure contract: missing API key, unreachable model,
  malformed JSON — all return [] and log an error. Never raises
  into the caller. The capture audit trail must not break on an
  optional side effect.
- Parser tolerates markdown fences, surrounding prose, invalid
  memory types, clamps confidence to [0,1], drops empty content.
- System prompt explicitly tells the model to return [] for most
  conversational turns (durable-fact bar, not "extract everything").
- Trust rules unchanged: candidates are never auto-promoted,
  extraction stays off the capture hot path, human triages via the
  existing CLI.

scripts/extractor_eval.py: new --mode {rule,llm} flag so the same
labeled corpus can be scored against both extractors. Default
remains rule so existing invocations are unchanged.

tests/test_extractor_llm.py: 12 new unit tests covering the parser
(empty array, malformed JSON, markdown fences, surrounding prose,
invalid types, empty content, confidence clamping, version tagging),
plus contract tests for missing API key, empty response, and a
mocked api_error path so failure modes never raise.

Test count: 264 -> 276 passing. No existing tests changed.

Next step: run `python scripts/extractor_eval.py --mode llm` against
the labeled set with ANTHROPIC_API_KEY in env, record the delta,
decide whether to wire LLM mode into the API endpoint and CLI or
keep it script-only for now.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-04-11 15:18:30 -04:00
7d8d599030 feat(eval-loop): Day 1+2 — labeled extractor corpus + baseline scorecard
Day 1 (labeled corpus):
- scripts/eval_data/interactions_snapshot_2026-04-11.json — frozen
  snapshot of 64 real claude-code interactions pulled from live
  Dalidou (test-client captures filtered out). This is the stable
  corpus the whole mini-phase labels against, independent of future
  captures.
- scripts/eval_data/extractor_labels_2026-04-11.json — 20 hand-labeled
  interactions drawn by length-stratified random sample. Positives:
  5/20 = ~25%, total expected candidates: 7. Plan deviation: Codex's
  plan asked for 30 (10/10/10 buckets); the real corpus is heavily
  skewed toward instructional/status content, so honest labeling of
  20 already crosses the fail-early threshold of "at least 5 plausible
  positives" without padding.

Day 2 (baseline measurement):
- scripts/extractor_eval.py — file-based eval runner that loads the
  snapshot + labels, runs extract_candidates_from_interaction on each,
  and reports yield / recall / precision / miss-class breakdown.
  Returns exit 1 on any false positive or false negative.

Current rule extractor against the labeled set:

    labeled=20  exact_match=15  positive_expected=5
    yield=0.0   recall=0.0     precision=0.0
    false_negatives=5           false_positives=0
    miss_classes:
      recommendation_prose
      architectural_change_summary
      spec_update_announcement
      layered_recommendation
      alignment_assertion

Interpretation: the rule-based extractor matches exactly zero of the
5 plausible positive interactions in the labeled set, and the misses
are spread across 5 distinct cue classes with no single dominant
pattern. This is the Day 4 hard-stop signal landing on Day 2 — a
single rule expansion cannot close a 5-way miss, and widening rules
blindly will collapse precision. The right move is to go straight to
the Day 4 decision gate and consider LLM-assisted extraction.

Escalating to DEV-LEDGER.md as R5 for human ratification before
continuing. Not skipping Day 3 silently.

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-04-11 15:11:33 -04:00