product guideMar 16, 2026·12 min read

How Linear Sprint Risk Analyzer Scores Velocity Deviation

By Jonathan Stocco, Founder

The Problem

It is Thursday afternoon and your engineering manager needs a velocity report for tomorrow’s leadership sync. She opens Linear, Notion, Slack, exports the last 3 sprints, calculates completion rates in a spreadsheet, compares against the previous quarter, and writes a summary. Two hours later, the report is done — and it is already missing this week’s data.

The gap is not data availability — it is analysis throughput. Raw ticket counts and status boards do not answer the questions that matter: which risks are systemic, which bottlenecks recur, which patterns predict delivery delays. Linear Sprint Risk Analyzer automates the engineering intelligence workflow, converting raw Linear, Notion, Slack data into structured analysis without manual compilation.

INFO

Engineering leads typically spend 2–4 hours weekly compiling this analysis manually. Linear Sprint Risk Analyzer delivers the same output in seconds, freeing time for technical work instead of reporting.

What This Blueprint Does

Four Agents. Sprint Risk Scoring. Notion + Slack Delivery.

The Linear Sprint Risk Analyzer pipeline runs 4 agents in sequence. Fetcher pulls data from Linear and Notion and Slack, and Formatter delivers the output. Here is what happens at each stage and why it matters.

  • Fetcher (Code Only): Queries Linear GraphQL API for the active sprint cycle — all issues, assignees, states, estimates, labels, dependencies, and update history.
  • Assembler (Code Only): Computes four sprint risk metrics from fetched data: velocity deviation (burn rate vs commitment), blocked chain depth (dependency chain analysis), scope creep ratio (mid-sprint additions), and concentration risk (work distribution Gini coefficient).
  • Analyst (Tier 2 Classification): Scores each risk dimension 0-100, classifies overall sprint health (HEALTHY/AT_RISK/CRITICAL), assigns per-issue risk levels (HIGH/MEDIUM/LOW), and generates ranked mitigation recommendations.
  • Formatter (Tier 3 Creative): Generates two outputs: (1) Notion sprint risk brief page with dimension breakdowns, per-issue risk table, blocked chain visualization, and mitigation recommendations.

When the pipeline completes, you get structured output that is ready to act on. The blueprint bundle includes everything needed to deploy, configure, and customize the workflow:

  • ITP-tested 24+3 node n8n workflow — import and deploy
  • Weekly schedule: fires every Monday at 9:00 UTC (customizable)
  • Four sprint risk dimensions: velocity deviation, blocked chains, scope creep, concentration risk
  • Sprint health classification: HEALTHY / AT_RISK / CRITICAL
  • Per-issue risk levels: HIGH / MEDIUM / LOW with specific risk factors
  • Mitigation recommendations ranked by priority and urgency
  • Notion sprint risk brief with dimension breakdowns and issue tables
  • Slack Block Kit digest with health status and top risks
  • Split-workflow pattern: scheduler + main pipeline (both included)
  • SINGLE-MODEL: the analysis model for analysis and formatting — no the primary reasoning modelneeded
  • AGGREGATE pattern: one Analyst call per weekly run, not per issue
  • ITP 8/8 variations, 14/14 milestones measured

Sprint window, metric calculations, and report format are configurable in the system prompts — adapt to your team’s workflow without modifying the pipeline. This means Linear Sprint Risk Analyzer adapts to your specific process, terminology, and integration requirements without forking the entire workflow.

TIP

All metric calculations and report formats are configurable in the system prompts. Adjust sprint windows, velocity baselines, and alert thresholds to match your team’s workflow.

How the Pipeline Works

Understanding how the pipeline works helps you customize it for your environment and troubleshoot issues when they arise. Here is a step-by-step walkthrough of the Linear Sprint Risk Analyzer execution flow.

Step 1: Fetcher

Tier: Code Only

The pipeline starts here. Queries Linear GraphQL API for the active sprint cycle — all issues, assignees, states, estimates, labels, dependencies, and update history. Pulls the complete sprint picture in a single API call. Zero LLM cost.

This stage ensures all downstream agents receive clean, validated input. If this step returns incomplete data, every downstream agent works with a degraded picture.

Step 2: Assembler

Tier: Code Only

Computes four sprint risk metrics from fetched data: velocity deviation (burn rate vs commitment), blocked chain depth (dependency chain analysis), scope creep ratio (mid-sprint additions), and concentration risk (work distribution Gini coefficient). Pure math, zero LLM cost.

Why this step matters: The result is a prioritized action queue, not just a data dump.

Step 3: Analyst

Tier: Tier 2 Classification

Scores each risk dimension 0-100, classifies overall sprint health (HEALTHY/AT_RISK/CRITICAL), assigns per-issue risk levels (HIGH/MEDIUM/LOW), and generates ranked mitigation recommendations. the analysis model with chain-of-thought enforcement.

Every field in the output is structured for the next agent to consume without parsing.

Step 4: Formatter

Tier: Tier 3 Creative

This is the final deliverable — what lands in your inbox or dashboard. Generates two outputs: (1) Notion sprint risk brief page with dimension breakdowns, per-issue risk table, blocked chain visualization, and mitigation recommendations. (2) Slack Block Kit digest with health status, dimension scores, top risks, and priority actions. the analysis model.

The entire pipeline executes without manual intervention. From trigger to output, every decision point follows a documented path. Every execution produces a traceable audit trail.

All nodes have been validated during Independent Test Protocol (ITP) testing on n8n v2.7.5. The error handling matrix in the bundle documents the recovery path for each failure mode.

INFO

This blueprint integrates with your existing Jira or Linear instance. No data leaves your infrastructure — all analysis runs in your own n8n environment.

Why we designed it this way

3 of 20 test deals had no activity history — no calls, no emails, no meetings. Without a dead letter queue, those 3 would have crashed the pipeline and blocked the other 17. The dead letter queue caught them; the pipeline processed the other 17 normally. Quarantine bad data, do not let it block good data.

— ForgeWorkflows Engineering

Cost Breakdown

Every metric is ITP-measured. The Linear Sprint Risk Analyzer scores sprint health across four risk dimensions — the analysis model for analysis and formatting, weekly aggregate cost.

The primary operating cost for Linear Sprint Risk Analyzer is the per-execution LLM inference cost. Based on Independent Test Protocol (ITP) testing, the measured cost is: Cost per Run: ~$0.04-0.07/week. This figure includes all API calls across all agents in the pipeline — not just the primary reasoning step, but every classification, scoring, and output generation call.

To put this in context, consider the manual alternative. A skilled team member performing the same work manually costs $60–90/hour for an engineering manager’s reporting time at a fully loaded rate (salary, benefits, tools, overhead). If the manual version of this workflow takes 2–4 hours weekly, the per-execution cost in human labor is significant. The blueprint executes the same pipeline for a fraction of that cost, with consistent quality and zero fatigue degradation.

Infrastructure costs are separate from per-execution LLM costs. You will need an n8n instance (self-hosted or cloud) and active accounts for the integrated services. The estimated monthly infrastructure cost is Weekly run cost ~$0.04-0.07/week ($0.16-$0.28/month), depending on your usage volume and plan tiers.

Quality assurance: Blueprint Quality Standard (BQS) audit result is 12/12 PASS. ITP result is all milestones PASS. These are not marketing claims — they are test results from structured inspection protocols that you can review in the product documentation.

All cost and performance figures are ITP-measured — tested against real data fixtures on n8n v2.7.5 in March 2026. See the product page for full test methodology.

TIP

Monthly projection: if you run this blueprint 100 times per month, multiply the per-execution cost by 100 and add your infrastructure costs. Most teams find the total is less than one hour of manual labor per month.

What's in the Bundle

9+ files — workflow JSON (main + scheduler), system prompts, and complete documentation.

When you purchase Linear Sprint Risk Analyzer, you receive a complete deployment bundle. This is not a SaaS subscription or a hosted service — it is a set of files that you own and run on your own infrastructure. Here is what is included:

  • linear_sprint_risk_analyzer_v1_0_0.json — The 24-node n8n main workflow (AGGREGATE weekly sprint analysis)
  • workflow/lsra_scheduler_v1_0_0.json — The 3-node scheduler workflow (Monday 9:00 UTC trigger)
  • README.md — 10-minute setup guide with Linear, Notion, Slack credentials and split-workflow configuration
  • docs/TDD.md — Technical Design Document with sprint risk taxonomy and SINGLE-MODEL pattern
  • system_prompts/analyst_system_prompt.md — Analyst prompt (4-dimension risk scoring + per-issue risk + mitigations)
  • system_prompts/formatter_system_prompt.md — Formatter prompt (Notion sprint risk brief + Slack Block Kit digest)
  • CHANGELOG.md — Version history

Start with the README.md. It walks through the deployment process step by step, from importing the workflow JSON into n8n to configuring credentials and running your first test execution. The dependency matrix lists every required service, API key, and estimated cost so you know exactly what you need before you start.

Every file in the bundle is designed to be read, understood, and modified. There is no obfuscated code, no compiled binaries, and no phone-home telemetry. You get the source, you own the source, and you control the execution environment.

Who This Is For

Linear Sprint Risk Analyzer is built for Engineering, Engineering Management teams that need to automate a specific workflow without building from scratch. If your team matches the following profile, this blueprint is designed for you:

  • You operate in a engineering or engineering management function and handle the workflow this blueprint automates on a recurring basis
  • You have (or are willing to set up) an n8n instance — self-hosted or cloud
  • You have active accounts for the required integrations: Linear account (Personal API key), Notion workspace (Integration with database access), Slack workspace (Bot Token with chat:write scope), Anthropic API key
  • You have API credentials available: Anthropic API, Linear API (httpHeaderAuth Bearer), Notion (httpHeaderAuth Bearer), Slack (Bot Token, httpHeaderAuth Bearer)
  • You are comfortable importing a workflow JSON and configuring API keys (the README guides you, but basic technical comfort is expected)

This is NOT for you if:

  • Does not create or modify Linear issues — it analyzes existing sprint data, not manages it
  • Does not replace sprint retrospectives — it provides quantitative risk inputs for your retro
  • Does not work with Jira, Asana, or other project tools — Linear GraphQL API only in v1.0
  • Does not predict individual issue completion dates — it scores systemic sprint-level risk
  • Does not guarantee sprint success — it provides early warning signals and mitigation recommendations
  • Does not summarize standups — that is what Slack Standup Summarizer does

Review the dependency matrix and prerequisites before purchasing. If you are unsure whether your environment meets the requirements, contact support@forgeworkflows.com before buying.

NOTE

All sales are final after download. Review the full dependency matrix, prerequisites, and integration requirements on the product page before purchasing. Questions? Contact support@forgeworkflows.com.

Edge cases to know about

Every pipeline has boundaries. These are intentional design decisions, not oversights — understanding them helps you deploy with the right expectations and plan for edge cases in your environment.

Does not create or modify Linear issues — it analyzes existing sprint data, not manages it

This is intentional. We default to human-in-the-loop for actions that carry reputational or financial risk. Once your team has validated output accuracy over 20+ cycles, you can adjust the pipeline to auto-execute — the workflow JSON supports it, but the default is conservative.

Does not replace sprint retrospectives — it provides quantitative risk inputs for your retro

We scoped this boundary after ITP testing revealed inconsistent results when the pipeline attempted this. The agents handle what they handle well — extending beyond this scope requires custom prompt engineering specific to your data shape.

Does not work with Jira, Asana, or other project tools — Linear GraphQL API only in v1.0

This keeps the pipeline focused on a single workflow. Adding this capability would introduce branching logic that varies by organization, and the tradeoff between complexity and reliability was not worth it for a reusable blueprint. Fork the workflow JSON if your use case demands it.

INFO

The dead letter queue captures any records that fail processing. Check it after your first production run to validate data coverage.

Getting Started

Deployment follows a structured sequence. The Linear Sprint Risk Analyzer bundle is designed for the following tools: n8n, Anthropic API, Linear, Notion, Slack. Here is the recommended deployment path:

  1. Step 1: Import workflows and configure credentials. Import linear_sprint_risk_analyzer_v1_0_0.json (main) and lsra_scheduler_v1_0_0.json (scheduler) into n8n. Configure Linear API credential (httpHeaderAuth with Bearer prefix), Notion integration (httpHeaderAuth with Bearer prefix), Slack Bot Token (httpHeaderAuth with Bearer prefix, chat:write scope), and Anthropic API key following the README.
  2. Step 2: Configure Linear team and output destinations. Set your LINEAR_TEAM_ID in the scheduler payload. Create a Notion database with Name (title), Health (select), and Date properties. Share the database with your Notion integration. Set the SLACK_CHANNEL for sprint risk digests. Update the scheduler webhook URL to point to the main workflow.
  3. Step 3: Activate and verify. Enable both workflows in n8n. Send a test POST to the main workflow webhook URL with your configuration. Verify the sprint risk analysis appears in Notion and the compact digest posts to Slack. The scheduler will auto-trigger every Monday at 9:00 UTC.

Before running the pipeline on live data, execute a manual test run with sample input. This validates that all credentials are configured correctly, all API endpoints are reachable, and the output format matches your expectations. The README includes test data examples for this purpose.

Once the test run passes, you can configure the trigger for production use (scheduled, webhook, or event-driven — depending on the blueprint design). Monitor the first few production runs to confirm the pipeline handles real-world data as expected, then let it run.

For technical background on how ForgeWorkflows blueprints are built and tested, see the Blueprint Quality Standard (BQS) methodology and the Inspection and Test Plan (ITP) framework. These documents describe the quality gates every blueprint passes before listing.

Ready to deploy? View the Linear Sprint Risk Analyzer product page for full specifications, pricing, and purchase.

TIP

Run a manual test with sample data before switching to production triggers. This catches credential misconfigurations and API endpoint issues before they affect real workflows.

Frequently Asked Questions

What sprint data does it analyze?+

It queries your Linear team's active cycle via GraphQL API — all issues with state, assignee, estimate, labels, dependencies, and update history. The Assembler computes four risk metrics from this data: velocity deviation, blocked chain depth, scope creep ratio, and concentration risk. The system prompts are standalone text files — edit scoring thresholds and output formats without touching the workflow JSON.

What are the four risk dimensions?+

Velocity Deviation measures burn rate vs commitment (are you on track to finish?). Blocked Chain Depth finds dependency chains where issues block other issues. Scope Creep Ratio tracks mid-sprint additions vs original scope. Concentration Risk measures work distribution — is one person carrying too much?

How does it classify sprint health?+

Each dimension is scored 0-100 using defined rubrics. HEALTHY means all dimensions are below 40. AT_RISK means any dimension is 40-69. CRITICAL means any dimension is 70 or above. The classification drives urgency in the Slack digest.

What does the Notion brief contain?+

A multi-dimension sprint risk page with: sprint health header, 4-dimension score breakdown with assessments, per-issue risk table (top issues by risk level), blocked chain details, scope creep analysis, concentration analysis, and ranked mitigation recommendations with urgency tags. Review the error handling matrix in the bundle — it documents the recovery path for each failure mode.

How does the Slack digest differ from the Notion brief?+

The Slack digest is a compact Block Kit message — sprint health status, all 4 dimension scores on one line, top 3 risks with reasons, and priority actions. Designed for quick scanning. The Notion brief has the full analysis for deeper review. The ITP test results in the bundle show measured performance across edge cases, not just happy-path data.

Why only Sonnet instead of Opus?+

The Fetcher retrieves all sprint data from Linear API. The Assembler pre-computes all risk metrics (velocity, chains, scope creep, Gini coefficient). The Analyst scores risk from pre-computed numbers — classification-tier reasoning that Sonnet 4.6 handles accurately. No deep causal analysis required.

How does it relate to Feature Request Extractor?+

Different direction entirely. FRE creates and labels Linear issues from feature requests (writes to Linear). LSRA reads Linear sprint data and analyzes risk (reads from Linear). FRE is per-issue triage. LSRA is weekly aggregate sprint analysis.

Can I run it on-demand instead of weekly?+

Yes. Send a POST request to the main workflow webhook URL with your configuration (linear_team_id, notion_database_id, slack_channel). The scheduler is optional — you can trigger analysis whenever you want. The README walks through configuration in under 10 minutes, including test data for validation.

Does it use web scraping?+

No. All data comes from the Linear GraphQL API. No web_search or external scraping. This makes the workflow fully deterministic and fast.

Is there a refund policy?+

All sales are final after download. Review the Blueprint Dependency Matrix and prerequisites before purchase. Questions? Contact support@forgeworkflows.com before buying. Full terms at forgeworkflows.com/legal.

How do I adjust the scoring thresholds for my team's workflow?+

All scoring parameters — velocity baselines, risk weights, and alert thresholds — are configurable in the system prompts. Open the relevant prompt file, adjust the threshold values, and re-run. No workflow JSON changes needed. The README includes a threshold tuning guide with recommended starting values.

Get Linear Sprint Risk Analyzer

$199

View Blueprint

Related Blueprints

Related Articles

Linear Sprint Risk Analyzer$199