Skip to content

AX: Agent Experience initiative — tracking issue #1068

@manwithacat

Description

@manwithacat

Tracking issue — Agent Experience (AX) initiative

Following the v0.67.118–125 high-velocity sprint, we have a structured way to discover AX bottlenecks: scripts/stall_log_mine.py (74a6b6d) mines Claude Code JSONL transcripts for confusion signals and produces a friction-ranked report.

This issue tracks the AX programme as a whole.

Programme goals

  1. Reduce agent friction on the top-N hot zones by ~50% measured by re-running stall_log_mine.py quarterly.
  2. Maintain the HTMX/Alpine/Fragment substrate as the empirical baseline. Stack-comparison work is deferred; longitudinal "this codebase vs itself last quarter" tells us if AX work is moving the number.
  3. Promote successful patterns into CLAUDE.md as durable guidance.

Child issues

All children closed ✓

5 of 5 closed across 27 versions (v0.67.118 → v0.67.144). The two large decompositions (#1064 + #1065) established the canonical mixin-per-family pattern — two reference implementations now exist, one with dict-based dispatch (region_adapter) and one with match-based dispatch (renderer).

Validation

Next monthly scripts/stall_log_mine.py sweep will quantify the friction-score impact. Expected changes vs. April baseline:

Issue can be closed once that sweep confirms the predictions. Until then keep open as the measurement anchor.

Measurement protocol

  • Re-run python scripts/stall_log_mine.py --days 30 at end of each month.
  • Diff the friction-ranked top-15 month-over-month.
  • A row dropping out of the top-15 = confirmed win. A row climbing = regression or new touch-zone.

Limitations of the current data

  1. Sample bias: window dominated by recent high-velocity work. Patterns reflect what was being worked on.
  2. Heuristic weights: repeat_reads + 3·edit_failures + 5·errors is a guess.
  3. Doesn't catch all stalls: agents giving up and asking the user, or stalling on a non-tool reasoning loop, leave no signal in the transcript tool stream.

Metadata

Metadata

Assignees

No one assigned

    Labels

    agent-experienceImprovements to codebase intelligibility/discoverability for AI coding agents

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions