Transformation Dossier Template

Type: Execution Recipe Confidence: 0.85 Sources: 7 Verified: 2026-03-31

Purpose

This recipe generates a personalized, evidence-based dossier for retailers showing digital transformation gaps — outdated tech stacks, absent AI commerce capabilities, or competitive pressure from digitally advanced rivals. The output is a 2-page document plus optional AI Readiness Scorecard appendix that names specific technology gaps with dated evidence, quantifies business impact using Forrester/Gartner benchmarks, and offers 3 graduated action options — ready for delivery to the decision-maker identified during enrichment. Directly feeds the Quickborn AI Readiness Diagnostic for Retail engagement pipeline. [src1, src2]

Prerequisites

Constraints

Tool Selection Decision

Which path?
├── Output format = PDF AND review mode = auto-deliver
│   └── PATH A: Automated PDF — Claude API + PageSpeed API + WeasyPrint + email delivery
├── Output format = PDF AND review mode = human review
│   └── PATH B: Reviewed PDF — Claude API + PageSpeed API + WeasyPrint + review queue
├── Output format = HTML email AND review mode = auto-deliver
│   └── PATH C: Automated HTML — Claude API + PageSpeed API + HTML template + email API
└── Output format = Markdown AND review mode = human review
    └── PATH D: Draft Markdown — Claude API + PageSpeed API + Markdown output + review queue
PathToolsCostSpeedOutput Quality
A: Automated PDFClaude API + PageSpeed API + WeasyPrint$0.02-0.10/dossier5-10 minHigh — requires confidence threshold (score 8+)
B: Reviewed PDFClaude API + PageSpeed API + WeasyPrint$0.02-0.10/dossier10-20 minExcellent — human catches edge cases
C: Automated HTMLClaude API + PageSpeed API + HTML template$0.02-0.08/dossier3-7 minHigh — inline rendering, no attachment friction
D: Draft MarkdownClaude API + PageSpeed API$0.02-0.05/dossier3-5 minGood — fastest for review-heavy workflows

Execution Flow

Step 1: Assemble Input Data

Duration: 2-5 minutes per dossier · Tool: Python/Node.js script + Google PageSpeed Insights API

Collect all required inputs from the enrichment and detection pipeline outputs. Run Core Web Vitals checks on the target company's site and up to 3 competitor sites. Assemble into a single structured context object for the LLM. Required fields: company name, domain, revenue, decision-maker name and title, at least 2 signals with source, date, and data point, Core Web Vitals for target and at least 1 competitor.

Verify: All required fields populated — company name, at least 1 decision-maker, at least 2 signals with source+date+data_point, Core Web Vitals for target and at least 1 competitor. · If failed: If fewer than 2 signals, return to detection pipeline and lower threshold, or flag company for monitoring rather than outreach.

Step 2: Generate Executive Summary

Duration: 1-2 minutes · Tool: Claude/GPT API

Generate a 1-paragraph executive summary that names the technology gap, competitive context, and quantified customer experience impact. Structure: Name the competitive context (how many competitors have moved ahead) → state the specific technology gap with data → quantify the customer experience cost using industry benchmarks → close with the conversion impact.

Example output: “2 of your 3 direct competitors launched AI-powered product discovery in Q1 2026. Your website still runs Magento 1.x — end-of-life since 2020 — with keyword-only search and a 4.1-second largest contentful paint. Among the 156 retailers Forrester tracked, those without AI commerce saw 15-25% lower conversion than AI-enabled peers. Your Core Web Vitals fail all 3 Google metrics where competitors score 84-91. Here's the customer experience gap that creates.”

Verify: Summary contains at least 2 specific data points with dates. No promotional language detected. · If failed: Regenerate with stricter system prompt constraints if promotional language appears.

Step 3: Generate Evidence Pack

Duration: 3-5 minutes · Tool: Claude/GPT API

Generate 4 evidence sections, all falsifiable:

  1. Tech Stack Comparison — Table: target vs. each competitor across e-commerce platform, search technology, recommendation engine, personalization capability, and mobile experience.
  2. Core Web Vitals Scores — Table: target vs. competitors across LCP, CLS, FID, and Performance Score, with Google pass/fail thresholds. [src6]
  3. GEO Readiness Audit — How the target appears in AI-powered search results vs. competitors: structured data quality, content freshness, schema markup, AI-crawlable content structure.
  4. Leadership Hiring Signals — Observable AI/data talent investment (or absence): AI/ML roles posted, data engineering positions, CDO/CTO hiring patterns.

Verify: All 4 evidence sections present. Tech stack comparison has at least 1 competitor. Core Web Vitals include numeric scores. Verification paths are actionable. · If failed: Remove any section lacking verifiable data or flag as “unverified — requires manual confirmation.”

Step 4: Generate Impact Analysis

Duration: 1-2 minutes · Tool: Claude/GPT API

Quantify business impact using Forrester/Gartner benchmarks with full citation. Show the math: “At $220M revenue, 15-25% lower conversion on 40% e-commerce share = $13.2M-$22M annual impact.” State time horizon explicitly and include confidence qualifier.

Available benchmarks:

Verify: At least 1 dollar-denominated impact figure derived from company revenue + benchmark range. Full citation present. · If failed: If revenue data missing, state benchmark as percentage with note.

Step 5: Generate AI Readiness Scorecard

Duration: 2-3 minutes · Tool: Claude/GPT API

Score the target across 6 dimensions of AI readiness (1-5 each, based only on observable public signals):

  1. Data Infrastructure — modern data architecture, CDW/data lake evidence, analytics tool adoption
  2. Automation Maturity — marketing automation, API integrations, headless commerce
  3. Organizational Receptivity — CDO/CTO hiring, transformation press releases, board technology background
  4. Compliance Readiness — privacy policy maturity, cookie consent, data handling practices
  5. AI Commerce Capability — recommendation engines, visual search, chatbots, personalization, dynamic pricing
  6. Workforce Adaptation — AI/ML job postings, data science team size, training programs

Aggregate: Total X/30. Readiness tiers: Critical (6-12), Developing (13-18), Emerging (19-24), Advanced (25-30).

Verify: All 6 dimensions scored with evidence. Aggregate score calculated correctly. Readiness tier matches score range. · If failed: If more than 2 dimensions have “no observable evidence,” return to enrichment pipeline for deeper research.

Step 6: Generate Recommended Actions

Duration: 1 minute · Tool: Claude/GPT API

Generate 3 graduated options:

Verify: All 3 options present. Option 1 is genuinely no-cost. Option 2 states $20K and 4-6 week timeline. Option 3 cites a benchmark. No urgency language detected. · If failed: Strip urgency language and regenerate Option 3 only.

Step 7: Assemble and Format Dossier

Duration: 1-3 minutes · Tool: PDF generation (WeasyPrint/Puppeteer) or HTML template

Assemble generated sections into the final dossier format: Executive Summary + Evidence Pack on page 1, Impact Analysis + Recommended Actions on page 2, AI Readiness Scorecard in appendix. Include methodology note and benchmark references.

Output files:

Verify: PDF renders at 2 pages plus optional 1-page scorecard appendix. All personalization fields populated. All source citations present. AI Readiness Scorecard totals correctly. · If failed: Reduce evidence pack to top 2 signals and regenerate.

Output Schema

{
  "output_type": "transformation_dossier",
  "format": "PDF + JSON",
  "sections": [
    {"name": "executive_summary", "type": "string", "required": true},
    {"name": "evidence_pack", "type": "object", "required": true},
    {"name": "impact_analysis", "type": "string", "required": true},
    {"name": "recommended_actions", "type": "array", "required": true},
    {"name": "ai_readiness_scorecard", "type": "object", "required": true}
  ],
  "scorecard_schema": {
    "dimensions": [
      {"name": "data_infrastructure", "score_range": "1-5"},
      {"name": "automation_maturity", "score_range": "1-5"},
      {"name": "organizational_receptivity", "score_range": "1-5"},
      {"name": "compliance_readiness", "score_range": "1-5"},
      {"name": "ai_commerce_capability", "score_range": "1-5"},
      {"name": "workforce_adaptation", "score_range": "1-5"}
    ],
    "aggregate_score_range": "6-30",
    "readiness_tiers": {
      "critical": "6-12", "developing": "13-18",
      "emerging": "19-24", "advanced": "25-30"
    }
  },
  "personalization_fields": [
    {"name": "company_name", "type": "string", "required": true},
    {"name": "decision_maker_name", "type": "string", "required": true},
    {"name": "decision_maker_title", "type": "string", "required": true},
    {"name": "tech_stack_findings", "type": "object", "required": true},
    {"name": "core_web_vitals_scores", "type": "object", "required": true},
    {"name": "competitor_names", "type": "array", "required": true}
  ],
  "expected_page_count": "2 + optional scorecard appendix",
  "sort_order": "summary, evidence, impact, actions; scorecard in appendix",
  "deduplication_key": "company_name + generated_at"
}

Quality Benchmarks

Quality MetricMinimum AcceptableGoodExcellent
Falsifiability rate> 80% of claims have verifiable source> 90%100%
Personalization completenessAll required fields populated+ competitor-specific comparisons+ industry-specific benchmarks
Benchmark citation accuracy1 cited benchmark with source2-3 benchmarks with n-countsAll benchmarks with full citation
Tone complianceNo “we/our/schedule” detected+ no superlatives or urgency+ reads as independent analyst report
Page length2 pages + appendix or fewer1.5-2 pages + focused appendix1.5 pages + 1-page scorecard
Tech stack comparison depthTarget vs. 1 competitorTarget vs. 2-3 competitorsTarget vs. 3 competitors + benchmark
AI Readiness ScorecardAll 6 dimensions scoredAll scored with evidenceAll scored with evidence + gap analysis

If below minimum: Regenerate the failing section with stricter system prompt constraints. If falsifiability rate is below 80%, audit each claim against the input data and remove any that lack a specific source.

Error Handling

ErrorLikely CauseRecovery Action
LLM generates promotional languageSystem prompt not restrictive enoughAdd explicit negative examples: “NEVER write: 'we can help', 'our solution', 'schedule a demo'”
Impact analysis lacks dollar figuresRevenue data missing from enrichmentUse percentage-only impact with note; flag company for manual revenue lookup
PageSpeed API returns errorsRate limit exceeded or site unreachableUse cached results if available (<7 days old); use WebPageTest API as fallback
Tech stack detection incompleteBuiltWith/Wappalyzer returns limited dataSupplement with manual site inspection; note “partial tech stack”
AI Readiness Scorecard has >2 unscored dimensionsEnrichment data insufficientDo not generate dossier — return to enrichment pipeline for deeper research
PDF exceeds 2 pages (excluding appendix)Too many signals or verbose generationLimit evidence pack to top 2 signals by composite score; reduce executive summary to 3 sentences
Personalization fields contain placeholdersEnrichment data incompleteDo not generate dossier — return to enrichment pipeline to fill gaps
Competitor data unavailableNo comparable retailers identifiedGenerate without competitive comparison; use industry benchmarks instead
LLM refuses impact projectionsSafety filter triggered by financial predictionsReframe as “industry benchmark range” rather than “prediction”; cite study explicitly

Cost Breakdown

ComponentPer Dossier50 Dossiers/Month500 Dossiers/Month
LLM API (Claude Sonnet)$0.02-0.05$1-2.50$10-25
LLM API (Claude Opus)$0.05-0.10$2.50-5$25-50
PageSpeed Insights API$0$0$0
PDF generation (WeasyPrint)$0$0$0
Email delivery (SendGrid)$0$0 (up to 100/day)$15/mo
Total (Sonnet + free tools)$0.02-0.05$1-2.50$10-40

Anti-Patterns

Wrong: Leading with AI hype

Dossier opens with “AI is transforming retail and your company is falling behind” or “Every retailer needs AI to survive.” Result: recipient classifies the dossier as generic thought leadership spam. Diagnostic credibility is destroyed by unsubstantiated claims. [src2]

Correct: Lead with their specific technology gap

Open with observable signals the recipient will recognize (“Your website runs Magento 1.x — end-of-life since June 2020. Your LCP is 4.1 seconds against competitors averaging 1.8 seconds.”). The measured gap sells the conversation — not AI hype.

Wrong: Comparing against Silicon Valley leaders

Dossier compares a $220M home goods retailer against Amazon, Walmart, or Shopify. Result: the comparison feels absurd and unactionable. The executive knows they are not Amazon. [src4]

Correct: Compare against direct competitors and category peers

“HomeNest (your direct competitor, similar revenue) launched AI-powered product discovery in January 2026 and reports a PageSpeed score of 91 vs. your 32.” Direct competitors create urgency; industry titans create despair.

Wrong: Scoring AI readiness without evidence

Scorecard shows “Automation Maturity: 2/5” with justification “Most retailers of this size have limited automation.” This is an assumption, not evidence. Indistinguishable from hallucination. [src1]

Correct: Every score cites a specific observable signal

“Automation Maturity: 2/5. Evidence: BuiltWith detects no marketing automation platform, no API gateway, no headless CMS. Career page shows manual data entry roles (3 posted Q1 2026). Verification: Check BuiltWith profile for mapleandvine.com.”

Wrong: Generic “you need to transform” recommendation

Option 2 says “We recommend a comprehensive digital transformation.” No price, no duration, no deliverable, no scope. Reads like every other consulting pitch. [src3]

Correct: Specific scope, price, timeline, and deliverable

“Option 2: Full AI Readiness Diagnostic ($20K, 4-6 weeks). Deliverable: 40-page assessment across 6 dimensions, vendor-neutral technology recommendations, implementation roadmap with quarterly milestones, and ROI model.”

Wrong: Fear-mongering in do-nothing projection

“If you don't act now, your company will lose customers and eventually fail.” This is speculative, unfalsifiable, and reads as a threat. [src2]

Correct: Benchmark-based trajectory projection

“Among 200 retailers tracked by Gartner (Predicts 2026), those in the bottom quartile of digital maturity lost 3-5 points of market share annually to digitally advanced competitors. At $220M revenue, 3-5 points = $6.6M-$11M in revenue moving to competitors with better digital experiences.”

When This Matters

Use this recipe when the signal detection pipeline has identified a retailer showing digital transformation gaps — outdated tech stack, absent AI commerce capabilities, poor Core Web Vitals, or competitive pressure from digitally advanced rivals — and the enrichment pipeline has resolved the signals to a specific company with decision-maker contacts. This is the asset generation step between enrichment (upstream) and outreach delivery (downstream). Unlike the distress dossier (which addresses operational pain), the transformation dossier addresses opportunity cost: what the retailer is losing by not keeping pace with digital-first competitors. The dossier directly feeds the Quickborn AI Readiness Diagnostic for Retail engagement pipeline.

Related Units