Transformation Dossier Template
What is the auto-generated dossier template for retailers showing digital transformation readiness?
Purpose
This recipe generates a personalized, evidence-based dossier for retailers showing digital transformation gaps — outdated tech stacks, absent AI commerce capabilities, or competitive pressure from digitally advanced rivals. The output is a 2-page document plus optional AI Readiness Scorecard appendix that names specific technology gaps with dated evidence, quantifies business impact using Forrester/Gartner benchmarks, and offers 3 graduated action options — ready for delivery to the decision-maker identified during enrichment. Directly feeds the Quickborn AI Readiness Diagnostic for Retail engagement pipeline. [src1, src2]
Prerequisites
- Enriched company profile available from
signal-library/retail/enrichment-mapping/2026— Retail Signal Enrichment Mapping - Signal detection report with scored signals available from
signal-library/retail/detection-rules/2026— Retail Detection Rules - Decision-maker contact (name, title, email/LinkedIn) from enrichment output
- Claude or GPT API key — from Anthropic Console or OpenAI Platform
- Google PageSpeed Insights API key — from Google Developers (free)
- PDF generation tool — Puppeteer, WeasyPrint, or equivalent (free)
Constraints
- Every claim must be falsifiable. Each data point requires a specific source, date, and verifiable figure. A recipient must be able to fact-check any claim within 5 minutes. [src1]
- Diagnostic tone only. The dossier reads like a technology audit: observation, measurement, options. Never use phrases like “we can help you,” “our solution,” or “schedule a demo.” [src2]
- Maximum 2 pages plus optional scorecard appendix. Retail executives evaluating digital transformation are time-constrained; dossiers exceeding 2 pages see 40-60% lower read-through rates. [src3]
- All industry benchmarks must cite the study, year, and sample size. Unsourced benchmarks are indistinguishable from hallucination. [src4]
- All personalization fields must be populated from enrichment data. Never deliver a dossier with placeholder text. Partially personalized outreach performs worse than generic outreach. [src5]
Tool Selection Decision
Which path?
├── Output format = PDF AND review mode = auto-deliver
│ └── PATH A: Automated PDF — Claude API + PageSpeed API + WeasyPrint + email delivery
├── Output format = PDF AND review mode = human review
│ └── PATH B: Reviewed PDF — Claude API + PageSpeed API + WeasyPrint + review queue
├── Output format = HTML email AND review mode = auto-deliver
│ └── PATH C: Automated HTML — Claude API + PageSpeed API + HTML template + email API
└── Output format = Markdown AND review mode = human review
└── PATH D: Draft Markdown — Claude API + PageSpeed API + Markdown output + review queue
| Path | Tools | Cost | Speed | Output Quality |
|---|---|---|---|---|
| A: Automated PDF | Claude API + PageSpeed API + WeasyPrint | $0.02-0.10/dossier | 5-10 min | High — requires confidence threshold (score 8+) |
| B: Reviewed PDF | Claude API + PageSpeed API + WeasyPrint | $0.02-0.10/dossier | 10-20 min | Excellent — human catches edge cases |
| C: Automated HTML | Claude API + PageSpeed API + HTML template | $0.02-0.08/dossier | 3-7 min | High — inline rendering, no attachment friction |
| D: Draft Markdown | Claude API + PageSpeed API | $0.02-0.05/dossier | 3-5 min | Good — fastest for review-heavy workflows |
Execution Flow
Step 1: Assemble Input Data
Duration: 2-5 minutes per dossier · Tool: Python/Node.js script + Google PageSpeed Insights API
Collect all required inputs from the enrichment and detection pipeline outputs. Run Core Web Vitals checks on the target company's site and up to 3 competitor sites. Assemble into a single structured context object for the LLM. Required fields: company name, domain, revenue, decision-maker name and title, at least 2 signals with source, date, and data point, Core Web Vitals for target and at least 1 competitor.
Verify: All required fields populated — company name, at least 1 decision-maker, at least 2 signals with source+date+data_point, Core Web Vitals for target and at least 1 competitor. · If failed: If fewer than 2 signals, return to detection pipeline and lower threshold, or flag company for monitoring rather than outreach.
Step 2: Generate Executive Summary
Duration: 1-2 minutes · Tool: Claude/GPT API
Generate a 1-paragraph executive summary that names the technology gap, competitive context, and quantified customer experience impact. Structure: Name the competitive context (how many competitors have moved ahead) → state the specific technology gap with data → quantify the customer experience cost using industry benchmarks → close with the conversion impact.
Example output: “2 of your 3 direct competitors launched AI-powered product discovery in Q1 2026. Your website still runs Magento 1.x — end-of-life since 2020 — with keyword-only search and a 4.1-second largest contentful paint. Among the 156 retailers Forrester tracked, those without AI commerce saw 15-25% lower conversion than AI-enabled peers. Your Core Web Vitals fail all 3 Google metrics where competitors score 84-91. Here's the customer experience gap that creates.”
Verify: Summary contains at least 2 specific data points with dates. No promotional language detected. · If failed: Regenerate with stricter system prompt constraints if promotional language appears.
Step 3: Generate Evidence Pack
Duration: 3-5 minutes · Tool: Claude/GPT API
Generate 4 evidence sections, all falsifiable:
- Tech Stack Comparison — Table: target vs. each competitor across e-commerce platform, search technology, recommendation engine, personalization capability, and mobile experience.
- Core Web Vitals Scores — Table: target vs. competitors across LCP, CLS, FID, and Performance Score, with Google pass/fail thresholds. [src6]
- GEO Readiness Audit — How the target appears in AI-powered search results vs. competitors: structured data quality, content freshness, schema markup, AI-crawlable content structure.
- Leadership Hiring Signals — Observable AI/data talent investment (or absence): AI/ML roles posted, data engineering positions, CDO/CTO hiring patterns.
Verify: All 4 evidence sections present. Tech stack comparison has at least 1 competitor. Core Web Vitals include numeric scores. Verification paths are actionable. · If failed: Remove any section lacking verifiable data or flag as “unverified — requires manual confirmation.”
Step 4: Generate Impact Analysis
Duration: 1-2 minutes · Tool: Claude/GPT API
Quantify business impact using Forrester/Gartner benchmarks with full citation. Show the math: “At $220M revenue, 15-25% lower conversion on 40% e-commerce share = $13.2M-$22M annual impact.” State time horizon explicitly and include confidence qualifier.
Available benchmarks:
- AI commerce gap: 15-25% lower conversion without AI product discovery (Forrester Digital Commerce 2025, n=156) [src1]
- Core Web Vitals: Sites with poor CWV see 24% higher bounce rates and up to 15% lower conversion (Google/web.dev 2025, n=10,000) [src6]
- Digital laggards: Bottom quartile lose 3-5 points of market share annually (Gartner Predicts 2026, n=200) [src4]
- AI talent pipeline: No AI hiring in 12 months = 18-24 months to build initial capability (Forrester 2025, n=89) [src1]
Verify: At least 1 dollar-denominated impact figure derived from company revenue + benchmark range. Full citation present. · If failed: If revenue data missing, state benchmark as percentage with note.
Step 5: Generate AI Readiness Scorecard
Duration: 2-3 minutes · Tool: Claude/GPT API
Score the target across 6 dimensions of AI readiness (1-5 each, based only on observable public signals):
- Data Infrastructure — modern data architecture, CDW/data lake evidence, analytics tool adoption
- Automation Maturity — marketing automation, API integrations, headless commerce
- Organizational Receptivity — CDO/CTO hiring, transformation press releases, board technology background
- Compliance Readiness — privacy policy maturity, cookie consent, data handling practices
- AI Commerce Capability — recommendation engines, visual search, chatbots, personalization, dynamic pricing
- Workforce Adaptation — AI/ML job postings, data science team size, training programs
Aggregate: Total X/30. Readiness tiers: Critical (6-12), Developing (13-18), Emerging (19-24), Advanced (25-30).
Verify: All 6 dimensions scored with evidence. Aggregate score calculated correctly. Readiness tier matches score range. · If failed: If more than 2 dimensions have “no observable evidence,” return to enrichment pipeline for deeper research.
Step 6: Generate Recommended Actions
Duration: 1 minute · Tool: Claude/GPT API
Generate 3 graduated options:
- Option 1 (Low commitment): Free 30-minute AI readiness preview — verbal walkthrough of scorecard findings and what peer retailers have done. No obligation.
- Option 2 (Medium commitment): Full AI Readiness Diagnostic ($20K, 4-6 weeks). Deliverable: 40-page assessment across 6 dimensions, vendor-neutral technology recommendations, implementation roadmap with quarterly milestones, and ROI model.
- Option 3 (No action): Do-nothing trajectory modeling — factual projection of what happens if digital gap continues for 2-4 more quarters. Benchmark-based projection, not fear-mongering. [src4]
Verify: All 3 options present. Option 1 is genuinely no-cost. Option 2 states $20K and 4-6 week timeline. Option 3 cites a benchmark. No urgency language detected. · If failed: Strip urgency language and regenerate Option 3 only.
Step 7: Assemble and Format Dossier
Duration: 1-3 minutes · Tool: PDF generation (WeasyPrint/Puppeteer) or HTML template
Assemble generated sections into the final dossier format: Executive Summary + Evidence Pack on page 1, Impact Analysis + Recommended Actions on page 2, AI Readiness Scorecard in appendix. Include methodology note and benchmark references.
Output files:
transformation_dossier_{company_slug}_{timestamp}.pdf— Formatted 2-page dossier + scorecard appendix ready for deliverytransformation_dossier_{company_slug}_{timestamp}.json— Structured dossier data for programmatic processingdossier_qa_checklist_{company_slug}.json— Quality assurance results from verification checks
Verify: PDF renders at 2 pages plus optional 1-page scorecard appendix. All personalization fields populated. All source citations present. AI Readiness Scorecard totals correctly. · If failed: Reduce evidence pack to top 2 signals and regenerate.
Output Schema
{
"output_type": "transformation_dossier",
"format": "PDF + JSON",
"sections": [
{"name": "executive_summary", "type": "string", "required": true},
{"name": "evidence_pack", "type": "object", "required": true},
{"name": "impact_analysis", "type": "string", "required": true},
{"name": "recommended_actions", "type": "array", "required": true},
{"name": "ai_readiness_scorecard", "type": "object", "required": true}
],
"scorecard_schema": {
"dimensions": [
{"name": "data_infrastructure", "score_range": "1-5"},
{"name": "automation_maturity", "score_range": "1-5"},
{"name": "organizational_receptivity", "score_range": "1-5"},
{"name": "compliance_readiness", "score_range": "1-5"},
{"name": "ai_commerce_capability", "score_range": "1-5"},
{"name": "workforce_adaptation", "score_range": "1-5"}
],
"aggregate_score_range": "6-30",
"readiness_tiers": {
"critical": "6-12", "developing": "13-18",
"emerging": "19-24", "advanced": "25-30"
}
},
"personalization_fields": [
{"name": "company_name", "type": "string", "required": true},
{"name": "decision_maker_name", "type": "string", "required": true},
{"name": "decision_maker_title", "type": "string", "required": true},
{"name": "tech_stack_findings", "type": "object", "required": true},
{"name": "core_web_vitals_scores", "type": "object", "required": true},
{"name": "competitor_names", "type": "array", "required": true}
],
"expected_page_count": "2 + optional scorecard appendix",
"sort_order": "summary, evidence, impact, actions; scorecard in appendix",
"deduplication_key": "company_name + generated_at"
}
Quality Benchmarks
| Quality Metric | Minimum Acceptable | Good | Excellent |
|---|---|---|---|
| Falsifiability rate | > 80% of claims have verifiable source | > 90% | 100% |
| Personalization completeness | All required fields populated | + competitor-specific comparisons | + industry-specific benchmarks |
| Benchmark citation accuracy | 1 cited benchmark with source | 2-3 benchmarks with n-counts | All benchmarks with full citation |
| Tone compliance | No “we/our/schedule” detected | + no superlatives or urgency | + reads as independent analyst report |
| Page length | 2 pages + appendix or fewer | 1.5-2 pages + focused appendix | 1.5 pages + 1-page scorecard |
| Tech stack comparison depth | Target vs. 1 competitor | Target vs. 2-3 competitors | Target vs. 3 competitors + benchmark |
| AI Readiness Scorecard | All 6 dimensions scored | All scored with evidence | All scored with evidence + gap analysis |
If below minimum: Regenerate the failing section with stricter system prompt constraints. If falsifiability rate is below 80%, audit each claim against the input data and remove any that lack a specific source.
Error Handling
| Error | Likely Cause | Recovery Action |
|---|---|---|
| LLM generates promotional language | System prompt not restrictive enough | Add explicit negative examples: “NEVER write: 'we can help', 'our solution', 'schedule a demo'” |
| Impact analysis lacks dollar figures | Revenue data missing from enrichment | Use percentage-only impact with note; flag company for manual revenue lookup |
| PageSpeed API returns errors | Rate limit exceeded or site unreachable | Use cached results if available (<7 days old); use WebPageTest API as fallback |
| Tech stack detection incomplete | BuiltWith/Wappalyzer returns limited data | Supplement with manual site inspection; note “partial tech stack” |
| AI Readiness Scorecard has >2 unscored dimensions | Enrichment data insufficient | Do not generate dossier — return to enrichment pipeline for deeper research |
| PDF exceeds 2 pages (excluding appendix) | Too many signals or verbose generation | Limit evidence pack to top 2 signals by composite score; reduce executive summary to 3 sentences |
| Personalization fields contain placeholders | Enrichment data incomplete | Do not generate dossier — return to enrichment pipeline to fill gaps |
| Competitor data unavailable | No comparable retailers identified | Generate without competitive comparison; use industry benchmarks instead |
| LLM refuses impact projections | Safety filter triggered by financial predictions | Reframe as “industry benchmark range” rather than “prediction”; cite study explicitly |
Cost Breakdown
| Component | Per Dossier | 50 Dossiers/Month | 500 Dossiers/Month |
|---|---|---|---|
| LLM API (Claude Sonnet) | $0.02-0.05 | $1-2.50 | $10-25 |
| LLM API (Claude Opus) | $0.05-0.10 | $2.50-5 | $25-50 |
| PageSpeed Insights API | $0 | $0 | $0 |
| PDF generation (WeasyPrint) | $0 | $0 | $0 |
| Email delivery (SendGrid) | $0 | $0 (up to 100/day) | $15/mo |
| Total (Sonnet + free tools) | $0.02-0.05 | $1-2.50 | $10-40 |
Anti-Patterns
Wrong: Leading with AI hype
Dossier opens with “AI is transforming retail and your company is falling behind” or “Every retailer needs AI to survive.” Result: recipient classifies the dossier as generic thought leadership spam. Diagnostic credibility is destroyed by unsubstantiated claims. [src2]
Correct: Lead with their specific technology gap
Open with observable signals the recipient will recognize (“Your website runs Magento 1.x — end-of-life since June 2020. Your LCP is 4.1 seconds against competitors averaging 1.8 seconds.”). The measured gap sells the conversation — not AI hype.
Wrong: Comparing against Silicon Valley leaders
Dossier compares a $220M home goods retailer against Amazon, Walmart, or Shopify. Result: the comparison feels absurd and unactionable. The executive knows they are not Amazon. [src4]
Correct: Compare against direct competitors and category peers
“HomeNest (your direct competitor, similar revenue) launched AI-powered product discovery in January 2026 and reports a PageSpeed score of 91 vs. your 32.” Direct competitors create urgency; industry titans create despair.
Wrong: Scoring AI readiness without evidence
Scorecard shows “Automation Maturity: 2/5” with justification “Most retailers of this size have limited automation.” This is an assumption, not evidence. Indistinguishable from hallucination. [src1]
Correct: Every score cites a specific observable signal
“Automation Maturity: 2/5. Evidence: BuiltWith detects no marketing automation platform, no API gateway, no headless CMS. Career page shows manual data entry roles (3 posted Q1 2026). Verification: Check BuiltWith profile for mapleandvine.com.”
Wrong: Generic “you need to transform” recommendation
Option 2 says “We recommend a comprehensive digital transformation.” No price, no duration, no deliverable, no scope. Reads like every other consulting pitch. [src3]
Correct: Specific scope, price, timeline, and deliverable
“Option 2: Full AI Readiness Diagnostic ($20K, 4-6 weeks). Deliverable: 40-page assessment across 6 dimensions, vendor-neutral technology recommendations, implementation roadmap with quarterly milestones, and ROI model.”
Wrong: Fear-mongering in do-nothing projection
“If you don't act now, your company will lose customers and eventually fail.” This is speculative, unfalsifiable, and reads as a threat. [src2]
Correct: Benchmark-based trajectory projection
“Among 200 retailers tracked by Gartner (Predicts 2026), those in the bottom quartile of digital maturity lost 3-5 points of market share annually to digitally advanced competitors. At $220M revenue, 3-5 points = $6.6M-$11M in revenue moving to competitors with better digital experiences.”
When This Matters
Use this recipe when the signal detection pipeline has identified a retailer showing digital transformation gaps — outdated tech stack, absent AI commerce capabilities, poor Core Web Vitals, or competitive pressure from digitally advanced rivals — and the enrichment pipeline has resolved the signals to a specific company with decision-maker contacts. This is the asset generation step between enrichment (upstream) and outreach delivery (downstream). Unlike the distress dossier (which addresses operational pain), the transformation dossier addresses opportunity cost: what the retailer is losing by not keeping pace with digital-first competitors. The dossier directly feeds the Quickborn AI Readiness Diagnostic for Retail engagement pipeline.