This recipe generates a personalized, evidence-based dossier for retailers showing digital transformation gaps — outdated tech stacks, absent AI commerce capabilities, or competitive pressure from digitally advanced rivals. The output is a 2-page document plus optional AI Readiness Scorecard appendix that names specific technology gaps with dated evidence, quantifies business impact using Forrester/Gartner benchmarks, and offers 3 graduated action options — ready for delivery to the decision-maker identified during enrichment. Directly feeds the Quickborn AI Readiness Diagnostic for Retail engagement pipeline. [src1, src2]
signal-library/retail/enrichment-mapping/2026 — Retail Signal Enrichment Mappingsignal-library/retail/detection-rules/2026 — Retail Detection RulesWhich path?
├── Output format = PDF AND review mode = auto-deliver
│ └── PATH A: Automated PDF — Claude API + PageSpeed API + WeasyPrint + email delivery
├── Output format = PDF AND review mode = human review
│ └── PATH B: Reviewed PDF — Claude API + PageSpeed API + WeasyPrint + review queue
├── Output format = HTML email AND review mode = auto-deliver
│ └── PATH C: Automated HTML — Claude API + PageSpeed API + HTML template + email API
└── Output format = Markdown AND review mode = human review
└── PATH D: Draft Markdown — Claude API + PageSpeed API + Markdown output + review queue
| Path | Tools | Cost | Speed | Output Quality |
|---|---|---|---|---|
| A: Automated PDF | Claude API + PageSpeed API + WeasyPrint | $0.02-0.10/dossier | 5-10 min | High — requires confidence threshold (score 8+) |
| B: Reviewed PDF | Claude API + PageSpeed API + WeasyPrint | $0.02-0.10/dossier | 10-20 min | Excellent — human catches edge cases |
| C: Automated HTML | Claude API + PageSpeed API + HTML template | $0.02-0.08/dossier | 3-7 min | High — inline rendering, no attachment friction |
| D: Draft Markdown | Claude API + PageSpeed API | $0.02-0.05/dossier | 3-5 min | Good — fastest for review-heavy workflows |
Duration: 2-5 minutes per dossier · Tool: Python/Node.js script + Google PageSpeed Insights API
Collect all required inputs from the enrichment and detection pipeline outputs. Run Core Web Vitals checks on the target company's site and up to 3 competitor sites. Assemble into a single structured context object for the LLM. Required fields: company name, domain, revenue, decision-maker name and title, at least 2 signals with source, date, and data point, Core Web Vitals for target and at least 1 competitor.
Verify: All required fields populated — company name, at least 1 decision-maker, at least 2 signals with source+date+data_point, Core Web Vitals for target and at least 1 competitor. · If failed: If fewer than 2 signals, return to detection pipeline and lower threshold, or flag company for monitoring rather than outreach.
Duration: 1-2 minutes · Tool: Claude/GPT API
Generate a 1-paragraph executive summary that names the technology gap, competitive context, and quantified customer experience impact. Structure: Name the competitive context (how many competitors have moved ahead) → state the specific technology gap with data → quantify the customer experience cost using industry benchmarks → close with the conversion impact.
Example output: “2 of your 3 direct competitors launched AI-powered product discovery in Q1 2026. Your website still runs Magento 1.x — end-of-life since 2020 — with keyword-only search and a 4.1-second largest contentful paint. Among the 156 retailers Forrester tracked, those without AI commerce saw 15-25% lower conversion than AI-enabled peers. Your Core Web Vitals fail all 3 Google metrics where competitors score 84-91. Here's the customer experience gap that creates.”
Verify: Summary contains at least 2 specific data points with dates. No promotional language detected. · If failed: Regenerate with stricter system prompt constraints if promotional language appears.
Duration: 3-5 minutes · Tool: Claude/GPT API
Generate 4 evidence sections, all falsifiable:
Verify: All 4 evidence sections present. Tech stack comparison has at least 1 competitor. Core Web Vitals include numeric scores. Verification paths are actionable. · If failed: Remove any section lacking verifiable data or flag as “unverified — requires manual confirmation.”
Duration: 1-2 minutes · Tool: Claude/GPT API
Quantify business impact using Forrester/Gartner benchmarks with full citation. Show the math: “At $220M revenue, 15-25% lower conversion on 40% e-commerce share = $13.2M-$22M annual impact.” State time horizon explicitly and include confidence qualifier.
Available benchmarks:
Verify: At least 1 dollar-denominated impact figure derived from company revenue + benchmark range. Full citation present. · If failed: If revenue data missing, state benchmark as percentage with note.
Duration: 2-3 minutes · Tool: Claude/GPT API
Score the target across 6 dimensions of AI readiness (1-5 each, based only on observable public signals):
Aggregate: Total X/30. Readiness tiers: Critical (6-12), Developing (13-18), Emerging (19-24), Advanced (25-30).
Verify: All 6 dimensions scored with evidence. Aggregate score calculated correctly. Readiness tier matches score range. · If failed: If more than 2 dimensions have “no observable evidence,” return to enrichment pipeline for deeper research.
Duration: 1 minute · Tool: Claude/GPT API
Generate 3 graduated options:
Verify: All 3 options present. Option 1 is genuinely no-cost. Option 2 states $20K and 4-6 week timeline. Option 3 cites a benchmark. No urgency language detected. · If failed: Strip urgency language and regenerate Option 3 only.
Duration: 1-3 minutes · Tool: PDF generation (WeasyPrint/Puppeteer) or HTML template
Assemble generated sections into the final dossier format: Executive Summary + Evidence Pack on page 1, Impact Analysis + Recommended Actions on page 2, AI Readiness Scorecard in appendix. Include methodology note and benchmark references.
Output files:
transformation_dossier_{company_slug}_{timestamp}.pdf — Formatted 2-page dossier + scorecard appendix ready for deliverytransformation_dossier_{company_slug}_{timestamp}.json — Structured dossier data for programmatic processingdossier_qa_checklist_{company_slug}.json — Quality assurance results from verification checksVerify: PDF renders at 2 pages plus optional 1-page scorecard appendix. All personalization fields populated. All source citations present. AI Readiness Scorecard totals correctly. · If failed: Reduce evidence pack to top 2 signals and regenerate.
{
"output_type": "transformation_dossier",
"format": "PDF + JSON",
"sections": [
{"name": "executive_summary", "type": "string", "required": true},
{"name": "evidence_pack", "type": "object", "required": true},
{"name": "impact_analysis", "type": "string", "required": true},
{"name": "recommended_actions", "type": "array", "required": true},
{"name": "ai_readiness_scorecard", "type": "object", "required": true}
],
"scorecard_schema": {
"dimensions": [
{"name": "data_infrastructure", "score_range": "1-5"},
{"name": "automation_maturity", "score_range": "1-5"},
{"name": "organizational_receptivity", "score_range": "1-5"},
{"name": "compliance_readiness", "score_range": "1-5"},
{"name": "ai_commerce_capability", "score_range": "1-5"},
{"name": "workforce_adaptation", "score_range": "1-5"}
],
"aggregate_score_range": "6-30",
"readiness_tiers": {
"critical": "6-12", "developing": "13-18",
"emerging": "19-24", "advanced": "25-30"
}
},
"personalization_fields": [
{"name": "company_name", "type": "string", "required": true},
{"name": "decision_maker_name", "type": "string", "required": true},
{"name": "decision_maker_title", "type": "string", "required": true},
{"name": "tech_stack_findings", "type": "object", "required": true},
{"name": "core_web_vitals_scores", "type": "object", "required": true},
{"name": "competitor_names", "type": "array", "required": true}
],
"expected_page_count": "2 + optional scorecard appendix",
"sort_order": "summary, evidence, impact, actions; scorecard in appendix",
"deduplication_key": "company_name + generated_at"
}
| Quality Metric | Minimum Acceptable | Good | Excellent |
|---|---|---|---|
| Falsifiability rate | > 80% of claims have verifiable source | > 90% | 100% |
| Personalization completeness | All required fields populated | + competitor-specific comparisons | + industry-specific benchmarks |
| Benchmark citation accuracy | 1 cited benchmark with source | 2-3 benchmarks with n-counts | All benchmarks with full citation |
| Tone compliance | No “we/our/schedule” detected | + no superlatives or urgency | + reads as independent analyst report |
| Page length | 2 pages + appendix or fewer | 1.5-2 pages + focused appendix | 1.5 pages + 1-page scorecard |
| Tech stack comparison depth | Target vs. 1 competitor | Target vs. 2-3 competitors | Target vs. 3 competitors + benchmark |
| AI Readiness Scorecard | All 6 dimensions scored | All scored with evidence | All scored with evidence + gap analysis |
If below minimum: Regenerate the failing section with stricter system prompt constraints. If falsifiability rate is below 80%, audit each claim against the input data and remove any that lack a specific source.
| Error | Likely Cause | Recovery Action |
|---|---|---|
| LLM generates promotional language | System prompt not restrictive enough | Add explicit negative examples: “NEVER write: 'we can help', 'our solution', 'schedule a demo'” |
| Impact analysis lacks dollar figures | Revenue data missing from enrichment | Use percentage-only impact with note; flag company for manual revenue lookup |
| PageSpeed API returns errors | Rate limit exceeded or site unreachable | Use cached results if available (<7 days old); use WebPageTest API as fallback |
| Tech stack detection incomplete | BuiltWith/Wappalyzer returns limited data | Supplement with manual site inspection; note “partial tech stack” |
| AI Readiness Scorecard has >2 unscored dimensions | Enrichment data insufficient | Do not generate dossier — return to enrichment pipeline for deeper research |
| PDF exceeds 2 pages (excluding appendix) | Too many signals or verbose generation | Limit evidence pack to top 2 signals by composite score; reduce executive summary to 3 sentences |
| Personalization fields contain placeholders | Enrichment data incomplete | Do not generate dossier — return to enrichment pipeline to fill gaps |
| Competitor data unavailable | No comparable retailers identified | Generate without competitive comparison; use industry benchmarks instead |
| LLM refuses impact projections | Safety filter triggered by financial predictions | Reframe as “industry benchmark range” rather than “prediction”; cite study explicitly |
| Component | Per Dossier | 50 Dossiers/Month | 500 Dossiers/Month |
|---|---|---|---|
| LLM API (Claude Sonnet) | $0.02-0.05 | $1-2.50 | $10-25 |
| LLM API (Claude Opus) | $0.05-0.10 | $2.50-5 | $25-50 |
| PageSpeed Insights API | $0 | $0 | $0 |
| PDF generation (WeasyPrint) | $0 | $0 | $0 |
| Email delivery (SendGrid) | $0 | $0 (up to 100/day) | $15/mo |
| Total (Sonnet + free tools) | $0.02-0.05 | $1-2.50 | $10-40 |
Dossier opens with “AI is transforming retail and your company is falling behind” or “Every retailer needs AI to survive.” Result: recipient classifies the dossier as generic thought leadership spam. Diagnostic credibility is destroyed by unsubstantiated claims. [src2]
Open with observable signals the recipient will recognize (“Your website runs Magento 1.x — end-of-life since June 2020. Your LCP is 4.1 seconds against competitors averaging 1.8 seconds.”). The measured gap sells the conversation — not AI hype.
Dossier compares a $220M home goods retailer against Amazon, Walmart, or Shopify. Result: the comparison feels absurd and unactionable. The executive knows they are not Amazon. [src4]
“HomeNest (your direct competitor, similar revenue) launched AI-powered product discovery in January 2026 and reports a PageSpeed score of 91 vs. your 32.” Direct competitors create urgency; industry titans create despair.
Scorecard shows “Automation Maturity: 2/5” with justification “Most retailers of this size have limited automation.” This is an assumption, not evidence. Indistinguishable from hallucination. [src1]
“Automation Maturity: 2/5. Evidence: BuiltWith detects no marketing automation platform, no API gateway, no headless CMS. Career page shows manual data entry roles (3 posted Q1 2026). Verification: Check BuiltWith profile for mapleandvine.com.”
Option 2 says “We recommend a comprehensive digital transformation.” No price, no duration, no deliverable, no scope. Reads like every other consulting pitch. [src3]
“Option 2: Full AI Readiness Diagnostic ($20K, 4-6 weeks). Deliverable: 40-page assessment across 6 dimensions, vendor-neutral technology recommendations, implementation roadmap with quarterly milestones, and ROI model.”
“If you don't act now, your company will lose customers and eventually fail.” This is speculative, unfalsifiable, and reads as a threat. [src2]
“Among 200 retailers tracked by Gartner (Predicts 2026), those in the bottom quartile of digital maturity lost 3-5 points of market share annually to digitally advanced competitors. At $220M revenue, 3-5 points = $6.6M-$11M in revenue moving to competitors with better digital experiences.”
Use this recipe when the signal detection pipeline has identified a retailer showing digital transformation gaps — outdated tech stack, absent AI commerce capabilities, poor Core Web Vitals, or competitive pressure from digitally advanced rivals — and the enrichment pipeline has resolved the signals to a specific company with decision-maker contacts. This is the asset generation step between enrichment (upstream) and outreach delivery (downstream). Unlike the distress dossier (which addresses operational pain), the transformation dossier addresses opportunity cost: what the retailer is losing by not keeping pace with digital-first competitors. The dossier directly feeds the Quickborn AI Readiness Diagnostic for Retail engagement pipeline.