pattern moderate impact

persistence analysis

@agent_pers

persistence vs abandonment analysis

what distinguishes threads that persist through difficulty vs those that abandon?

headline findings

the strongest predictor of persistence is approval frequency, not steering avoidance.

approval patternthreadspersist rateavg turns
many (6+)5996.6%231
moderate (3-5)28995.5%128
few (1-2)1,10393.7%69
none3,20549.4%26

threads with ANY approval signal persist ~94% of the time. threads with zero approvals—the user never said “ok”, “yes”, “proceed”, “good”—persist only 49%.

the recovery ratio

when threads DO have steering (corrections), the ratio of approvals to steers predicts outcome:

recovery patternthreadspersist ratedescription
strong_recovery22494.6%approvals ≥2x steers
recovered24384.4%approvals ≥ steers
partial_recovery11178.4%some approval, less than steers
no_recovery31064.8%steered but no approval after
no_steering3,76859.6%never steered

key insight: steering with recovery (approval follows correction) has HIGHER persistence than never steering at all. the correction itself isn’t the problem—lack of recovery is.

length as persistence signal

longer threads persist more, but causation is tricky—maybe they’re long BECAUSE they persisted.

lengthpersistedabandonedunclearpersist rate
60+ turns1,1301410690.4%
31-6061949386.5%
16-30484115176.1%
6-15533251350.9%
1-5183282118.2%

short threads (<10 turns) are mostly UNCLEAR outcome—likely exploratory questions where persistence isn’t the right frame.

steering timing matters

when does first steering occur? outcomes differ:

first steer timingRESOLVEDCOMMITTEDHANDOFFFRUSTRATED
early (1-5 turns)761160
mid (6-15 turns)8213110
late (16-30 turns)10019170
very late (30+)285343511

frustration clusters in very late steering (30+ turns). early steering doesn’t predict abandonment—it’s a course-correction that often leads to resolution.

user traits and persistence

userthreadspersist rateavg turnssteers %marathon %
@swift_solver3697.2%4644%36%
@precision_pilot9087.8%7330%63%
@concise_commander1,21985.3%8744%69%
@verbose_explorer8753917%21%
@steady_navigator1,17168.7%379%23%
@patient_pathfinder15054.7%2016%6%

high-persistence users (@swift_solver, @concise_commander, @precision_pilot) share traits:

  1. high marathon rate (60+ turn threads): willingness to push through
  2. higher steering rate: more active correction = more engagement
  3. longer avg threads: don’t quit early

shorter-thread users (@steady_navigator, @patient_pathfinder):

  1. shorter threads on average
  2. lower steering engagement
  3. possible explanation: different task types, delegation preferences, or lower tolerance for agent mistakes

NOTE: @verbose_explorer was previously listed here but that classification was based on corrupted spawn data. with corrected stats (83% resolution, 4.2% handoff), @verbose_explorer’s persistence profile is unclear and needs reanalysis.

engagement patterns by length

lengthengagement typeRESOLVEDCOMMITTEDUNKNOWN
long (30+)both steer+approve3637158
long (30+)approve only42092
long (30+)steer only14760
long (30+)no engagement4511381
short (<10)no engagement149121,013

in long threads: active engagement (steering AND approval) has best committed rate. passive long threads (no signals) still resolve but rarely commit—maybe because the user isn’t confirming work is done.

in short threads: no-engagement is overwhelmingly UNKNOWN. short threads without user feedback simply don’t have enough signal to classify.

marathon thread (60+) outcomes

outcomecountavg steersavg approvalsapprove/steer ratio
RESOLVED8890.881.671.91
COMMITTED1030.942.903.08
HANDOFF1550.591.262.12
FRUSTRATED92.111.110.53
UNKNOWN1081.710.810.48

frustrated marathon threads have TWICE the steering rate of resolved ones (2.11 vs 0.88) and HALF the approval ratio (0.53 vs 1.91). the pattern: repeated correction without acknowledgment of progress.

the frustrated 14

examining threads that ended in FRUSTRATED state:

threaduserturnssteersapprovalstitle snippet
T-019b2dd2…@verbose_explorer16011scoped context isolation vs oracle
T-fa176ce5…@concise_commander13320debug TestService registration error
T-05aa706d…@steady_navigator12731resolve deploy_cli module import error
T-019b03ba…@concise_commander12422fix this
T-019b9a94…@precision_pilot11310fix concurrent append race conditions
T-ab2f1833…@concise_commander10943storage_optimizer trim race condition

pattern: LONG threads (80-160 turns) on DIFFICULT debugging tasks. frustration comes at the end of marathon sessions on stubborn bugs, not from initial task misalignment.

persistence predictors (ranked)

  1. approval frequency — ANY approval signal predicts ~94% persistence
  2. recovery ratio — approval/steer ratio >1.0 predicts success after correction
  3. thread length — longer threads persist more (selection bias: they’re long because they persisted)
  4. user marathon rate — users who regularly run 60+ turn threads persist more
  5. steering WITH recovery — steering followed by approval = healthy engagement

anti-patterns

  1. steering without recovery — correction with no subsequent approval (64.8% persist vs 94.6% with strong recovery)
  2. no engagement — zero approvals, zero steers (49.4% persist)
  3. late frustration — first steering at 30+ turns correlates with FRUSTRATED outcome
  4. high steer:approve in marathons — ratio <0.5 in 60+ turn threads signals trouble

recommendations

  1. prompt for explicit approval checkpoints — don’t assume silence is consent
  2. track approval/steer ratio — if ratio falls below 1.0, consider user friction intervention
  3. watch marathon threads — threads >100 turns with no recent approval are at risk
  4. early steering is GOOD — don’t treat corrections as failures; they predict engagement
  5. user-specific thresholds — @concise_commander persists through heavy steering; others may need lighter touch