Zum Inhalt springen

Skills / progressive estimation

progressive estimation

An AI skill for estimating AI-assisted and hybrid human+agent development work. Research-backed formulas. PERT statistics. Calibration feedback loops. Zero dependencies.

1von @Enreignvor 33d aktualisiertMITGitHub →

Installation

Kompatibilitaet

Claude CodeGeminiCursorVS Code

Beschreibung

Progressive Estimation

An AI skill for estimating AI-assisted and hybrid human+agent development work.

Research-backed formulas. PERT statistics. Calibration feedback loops. Zero dependencies.


[!WARNING] Early development. Progressive Estimation is actively developed and the formulas, multipliers, and default values change frequently. Expect rough edges, incomplete calibration, and breaking changes between versions. Bug reports, calibration data, and PRs are welcome.

[!NOTE] Estimation is one of the hardest problems in software engineering. A database migration might take 2 days for a small app or 2 years for a large enterprise system. Meanwhile, small tasks that once took hours can now be completed in minutes with modern AI tools. This skill gives you structure and consistency — not certainty. See DISCLAIMER.md for the full picture.


What It Does

  • Estimates development tasks accounting for both human and AI agent effort
  • Supports single tasks or batches (paste 5 issues or 500)
  • Produces PERT expected values with confidence bands, not just ranges
  • Separates "expected" from "committed" estimates at your chosen confidence level
  • Estimates token consumption and API cost per model tier (economy/standard/premium)
  • Outputs in formats ready for Linear, JIRA, ClickUp, GitHub Issues, Monday, GitLab, Asana, Azure DevOps, Zenhub, and Shortcut
  • Includes a calibration system to improve accuracy over time with actuals

Quick Start

Install

Claude Code (recommended — full progressive loading):

git clone https://github.com/Enreign/progressive-estimation.git ~/.claude/skills/progressive-estimation

Other clients: See the full Installation Guide for setup instructions for Cursor, GitHub Copilot, Windsurf, Cline, Aider, Continue.dev, ChatGPT, and Gemini Code Assist.

Tell Your Agents

Copy the example files into your project so agents know the skill is available and what it can do:

  • examples/CLAUDE.md — project-level instructions (features, conventions, customization)
  • examples/AGENTS.md — subagent instructions (when to estimate, how to invoke, re-estimation triggers)

Merge the relevant sections into your existing CLAUDE.md / AGENTS.md, or copy them as-is. Both files include customizable sections for your team's defaults.

Use

In your AI coding client, just ask for an estimate:

Estimate: "Add Stripe payment integration to our checkout flow"

Or batch estimate:

Estimate these tasks:
1. Add dark mode toggle
2. Migrate database from MySQL to PostgreSQL
3. Build Slack notification service
4. Implement CSV export for reports
5. Set up end-to-end test suite

The skill auto-triggers on keywords like estimate, how long, effort, sizing, story points.

[!TIP] Works with any AI coding client that supports custom instructions. Claude Code, Cursor, and GitHub Copilot get progressive loading (files load on demand). All others work with the full skill loaded at once. See INSTALLATION.md for your client.


How It Works

Progressive Disclosure

The skill asks only what it needs. Two paths, two scopes:

| Path | Questions | Best For | |------|-----------|----------| | Quick | 4 questions + defaults | Fast sizing, backlog grooming | | Detailed | 13 questions, full control | Sprint commitments, external deadlines |

Quick + Single    fastest, ~4 questions
Quick + Batch     paste a list, get a table
Detailed + Single full intake, rich output
Detailed + Batch  shared defaults + per-task overrides

Computation Pipeline

Agent Rounds x Minutes per Round
    + Integration Time
    + Human Fix Time (agent-effectiveness-adjusted)
    + Human Review Time
    + Human Planning Time
    x Org Size Overhead (human time only)
    x Task Type Multiplier
    -> Cone of Uncertainty spread
    -> PERT Expected + Standard Deviation
    -> Confidence Multiplier
    -> Token & Cost Estimation
    = Expected Estimate + Committed Estimate + Token Estimate

Key Concepts

Based on METR research (24k runs, 228 tasks, Jan 2025–Feb 2026): AI agents handle ~90% of small tasks well but only ~30% of XL tasks. Calibrated against METR autonomous completion rates with partial-credit adjustment.

| Size | Agent Effectiveness | Human Fix Adjustment | |------|-------------------|---------------------| | S | 90% | Minimal correction | | M | 50% | Significant intervention | | L | 35% | Mostly human-driven | | XL | 30% | Human-driven with agent assist |

Every estimate produces a weighted expected value using log-normal weighting (validated against 84k tasks via KS test — log-normal fits better than PERT-beta in all size bands):

Expected = (min + 4 x geometric_mean(min, max) + max) / 6
SD = (max - min) / 6

This gives stakeholders a single "most likely" number plus confidence bands (68%, 95%). The geometric mean shifts the expected value toward the minimum, reflecting the right-skewed nature of software effort distributions.

Separate "what we expect" from "what we commit to". Size-dependent multipliers derived from 84k estimate-actual pairs (small tasks need larger buffers due to wider variance):

| Level | S | M | L | XL | Use Case | |-------|---|---|---|----|----| | 50% | 1.0x | 1.0x | 1.0x | 0.75x | Stretch goal, internal planning | | 80% | 1.8x | 1.4x | 1.4x | 1.5x | Likely delivery (default) | | 90% | 2.9x | 2.1x | 2.0x | 2.2x | Safe commitment, external deadlines |

Calibrated from CESAW, SiP, Renzo, Project-22 datasets with bootstrap 95% CIs.

Early-phase estimates get wider ranges automatically:

| Phase | Spread | Accuracy | |-------|--------|----------| | Concept | 2.0x wider | Can be off by 2-4x | | Requirements | 1.5x wider | Major decisions made | | Design | 1.2x wider | Most unknowns resolved | | Ready to build | Baseline | Clear spec |

Based on Construx research.

Every estimate includes token consumption based on complexity × maturity lookup tables, split into input/output tokens. Optionally estimates API cost across three model tiers:

| Tier | Models | Input/1M | Output/1M | |------|--------|----------|-----------| | Economy | Haiku, GPT-4o Mini, Gemini Flash | $0.50 | $2.50 | | Standard | Sonnet, GPT-4o, Gemini 2.5 Pro | $2.50 | $12.00 | | Premium | Opus, GPT-5 | $5.00 | $25.00 |

Tokens appear in the one-line summary: 10-26 agent rounds (~180k tokens). Cost only appears when explicitly requested (show_cost=true).

Different work has different lifecycle overhead:

| Type | Multiplier | Why | |------|-----------|-----| | Coding | 1.0x | Baseline | | Bug fix | 1.3x | Debugging, reproduction, regression testing | | Investigation | 0.5x | Timeboxed — output is a plan, not code | | Design | 1.2x | Iteration with stakeholders | | Testing | 1.3x | Environment setup, fixtures, flakiness | | Infrastructure | 1.5x | Provisioning, CI/CD, deployment verification | | Data migration | 2.0x | Planning, validation, rollback, staged rollout |


Example Output

Single Task (Quick Mode)

Expected: ~4 hrs | Committed (80%): ~5.5 hrs | 10-26 agent rounds (~180k tokens) + 3 hrs human | Risk: medium | Size: M

PERT Expected: 4.2 hrs (most likely outcome)
Standard Deviation: +/-0.8 hrs
68% Confidence: 3.4 - 5.0 hrs
95% Confidence: 2.6 - 5.8 hrs

| Field | Value | |-------|-------| | Complexity | M | | Task Type | coding | | Agent Rounds | 10-26 | | Agent Time | 20-78 min | | Human Review | 30 min | | Human Planning | 30-60 min | | Human Fix/QA | 8-30 min | | Expected (PERT) | ~4 hrs | | Committed (80%) | ~5.5 hrs | | Token Estimate | ~180k tokens | | Model Tier | standard | | Risk | medium | | Team | 1 human, 1 agent |

Batch (Quick Mode)

5 tasks | Expected: ~23.5 hrs | Committed (80%): ~32.8 hrs | 2S, 2M, 1L

| # | Task | Size | Type | Expected | Committed | Risk | |---|------|------|------|----------|-----------|------| | 1 | Dark mode toggle | S | coding | ~1.3 hrs | ~1.8 hrs | low | | 2 | DB migration | L | data-mig | ~14.2 hrs | ~19.8 hrs | high | | 3 | Slack notifier | M | coding | ~2.9 hrs | ~4.1 hrs | med | | 4 | CSV export | S | coding | ~1.3 hrs | ~1.8 hrs | low | | 5 | E2E test suite | M | testing | ~3.8 hrs | ~5.3 hrs | med |

Warning: Task #2 is type=data-migration (2.0x overhead). Consider phased delivery.


Issue Tracker Integration

Estimates can be output in two modes for any supported tracker:

| Mode | How It Works | Setup | |------|-------------|-------| | Embedded (default) | Markdown table in description/body | None | | Native | Maps to tracker-specific fields | Custom fields |

Supported: Linear, JIRA, ClickUp, GitHub Issues, Monday, GitLab, Asana, Azure DevOps, Zenhub, Shortcut

Embedded mode works everywhere immediately. Native mode requires custom fields for agent-specific metrics.


Standalone Calculator

Don't want to depend on an LLM for arithmetic? Ask the skill to generate a deterministic calculator:

Generate an estimation calculator in Python

Generates a single-file, zero-dependency script from the canonical formulas. Accepts inputs via CLI args or stdin JSON, outputs the full estimate as JSON.

Supported languages: Python, TypeScript, JavaScript, Rust, Go, Ruby, Java, C#, Swift, Kotlin


Calibration

Estimates improve with feedback. The skill includes a calibration system:

| Step | What | |------|------| | 1. Log actuals | Record estimated vs. actual effort after completing work | | 2. Track PRED(25) | Percentage of estimates within 25% of actual (target: 75%) | | 3. Reference stories | Maintain examples per size per task type as anchors | | 4. Bias detection | Identify systematic over/under estimation | | 5. Team profiles | Separate calibration per team |

Most teams reach PRED(25) >= 65% within 3-5 calibration cycles and >= 75% within 8-12 cycles.

See references/calibration.md for the full system.


File Structure

progressive-estimation/
├── SKILL.md                    Workflow map (loaded first, always)
├── INSTALLATION.md             Setup guide for 9 AI coding clients
├── DISCLAIMER.md               Honest limitations of estimation
├── CONTRIBUTING.md             How to contribute
├── CODE_OF_CONDUCT.md          Community guidelines
├── README.md
├── LICENSE                     MIT
├── .github/
│   ├── PULL_REQUEST_TEMPLATE.md
│   └── ISSUE_TEMPLATE/
│       ├── bug_report.md
│       ├── feature_request.md
│       └── calibration_data.md
├── references/
│   ├── questionnaire.md        Progressive intake (phase 1)
│   ├── frameworks.md           Round-based, module/wave, S-M-L (phase 2)
│   ├── formulas.md             All arithmetic, single source of truth (phase 3)
│   ├── output-schema.md        Output formats, tracker mappings (phase 4)
│   └── calibration.md          Tuning with actuals (phase 5, on request)
├── tests/
│   ├── test_formulas.py        63 regression tests (deterministic, stdlib-only)
│   ├── validate_all_datasets.py  86k+ data point validation across 13 datasets
│   └── deep_validation.py      11-analysis deep validation with Parameter Audit Card
├── datasets/
│   ├── README.md               Download instructions for research datasets
│   ├── download_benchmarks.sh  Downloads METR, OpenHands, Aider benchmark data
│   └── benchmarks/             Bundled + downloaded benchmark datasets
├── docs/
│   └── deep-validation-findings.md  Combined findings from 110k data points
└── evals/
    ├── eval-quick.md           Quick path smoke test
    ├── eval-hybrid.md          Detailed path, multi-team
    ├── eval-batch.md           Batch with dependencies
    └── eval-regression.md      Known-good baselines

Files are loaded progressively — the skill only reads what it needs for the current phase. SKILL.md is the map; reference files are the territory.


Research

The estimation model is informed by:

| Source | Contribution | |--------|-------------| | METR | Agent effectiveness decay by task size; 24k runs across 228 tasks (arXiv:2503.14499) | | Derek Jones datasets | 86k+ tasks (CESAW, SiP, Project-22, Renzo) — confidence multipliers, distribution fitting, review times | | Aider leaderboard | ~50 models with cost, tokens, pass rates — output token ratios, cost model validation | | PERT / Log-Normal | Three-point estimation, validated with log-normal fit (KS test, n=84k) | | Jorgensen & Grimstad | Calibration feedback improving accuracy 64% -> 81% | | Construx | Cone of Uncertainty — estimate ranges narrowing as decisions are made | | Standish CHAOS | Project overrun patterns and their limitations |

Deep Validation

All formula parameters have been validated against 110k+ data points across 13 datasets using bootstrap CIs and KS goodness-of-fit tests. See docs/deep-validation-findings.md for the full audit (53 parameters, 11 analyses).


Evals

Evaluation prompts per the Claude Skills 2.0 framework:

| Eval | Tests | |------|-------| | eval-quick.md | Quick path produces valid PERT output with minimal input | | eval-hybrid.md | Detailed path handles multi-team, confidence levels, org overhead | | eval-batch.md | Batch mode with mixed types, dependencies, and rollup | | eval-regression.md | 8 baseline cases to detect drift after formula changes |

Run evals after any change to formulas, frameworks, or the skill workflow.


Contributing

Contributions welcome — see CONTRIBUTING.md for guidelines. Key areas:

  • Calibration data — Share anonymized estimated vs. actual results to improve default ratios
  • Tracker mappings — Additional tracker support (Notion, Basecamp, etc.)
  • Task types — New multipliers for work categories not yet covered
  • Formulas — Improvements backed by data or research
  • Evals — Additional test cases, especially edge cases

Please include research citations or empirical data when proposing formula changes.

This project follows a Code of Conduct.


License

MIT — Copyright (c) 2026 Stanislav Shymanskyi

Aehnliche Skills