This recipe produces a validated, revenue-maximizing price point — backed by Van Westendorp range analysis (acceptable price boundaries), Gabor-Granger demand curve modeling (revenue-optimal point), and optional live A/B experiment data (real conversion validation) — within 4-13 weeks at $0-$5,000. It outputs a price sensitivity report with intersection analysis, a segmented demand curve with revenue projections, and experiment results that feed directly into pricing page updates, sales enablement, and financial modeling. A 1% improvement in pricing yields an 11.1% increase in operating profit — making this the highest-leverage experiment a startup can run. [src2]
Which path?
├── Under 100 prospects AND budget = $0
│ └── PATH A: Qualitative — 5-10 pricing interviews + Google Forms Van Westendorp
├── 100-500 prospects AND budget = $0-$500
│ └── PATH B: DIY Survey — Google Forms + manual VW + Gabor-Granger
├── 100-2,000 prospects AND budget = $500-$2,000
│ └── PATH C: Automated — Conjointly analysis + optional cohort test
└── 2,000+ customers AND budget > $2,000
└── PATH D: Full Stack — Survey research + live A/B + statistical validation
| Path | Tools | Cost | Speed | Output Quality |
|---|---|---|---|---|
| A: Qualitative | Zoom, Google Forms, Sheets | $0 | 2-3 weeks | Directional — range only, no demand curve |
| B: DIY Survey | Google Forms, Sheets | $0-$100 | 4-6 weeks | Good — VW range + GG revenue curve |
| C: Automated | Conjointly, Sheets | $500-$2K | 4-8 weeks | High — auto analysis + segment breakdown |
| D: Full Stack | Conjointly, Stripe, PostHog | $2K-$5K | 8-13 weeks | Excellent — survey + live validation |
Duration: 2-4 hours · Tool: Google Sheets
Record current price, how it was set, current conversion rate, and ARPU. Formulate 2-3 hypotheses with specific numeric predictions. Define 5-8 test price points at 15-30% increments spanning 0.5x-3x current price. [src3]
Hypothesis template:
"We believe changing from $[current] to $[test] will [increase/decrease]
[metric] by [X]% because [evidence/reasoning]."
Price points (example, current = $29):
$19, $25, $29, $39, $49, $59, $79 (7 points, ~25% increments)
Verify: 2+ hypotheses documented; 5-8 price points defined · If failed: Analyze 5 competitor pricing pages and talk to 3 customers about value perception
Duration: 1-2 weeks · Tool: Google Forms or Conjointly (free tier)
Design survey with clear product description + four Van Westendorp questions: (1) too cheap to trust quality, (2) bargain/great deal, (3) expensive but would consider, (4) too expensive to consider. Distribute to 150-300 prospects. Plot cumulative frequency curves and find four intersection points: PMC, IPP, OPP, PME. The acceptable price range spans PMC to PME. [src1] [src5]
Van Westendorp intersection points:
PMC (Marginal Cheapness) = "Too Cheap" ∩ "Expensive" → price floor
IPP (Indifference) = "Bargain" ∩ "Expensive" → median acceptable
OPP (Optimal Price) = "Too Cheap" ∩ "Too Expensive" → fewest reject
PME (Marginal Expensive) = "Bargain" ∩ "Too Expensive" → price ceiling
Example: PMC=$22 | IPP=$35 | OPP=$31 | PME=$58
Acceptable range: $22-$58
Verify: 100+ responses; all 4 curves plotted; PMC-to-PME range identified · If failed: Add $10 incentive; extend collection 1 week; if curves never cross, sample too small [src5]
Duration: 1-2 weeks (can run parallel with VW using separate respondents) · Tool: Google Forms (branching) or Conjointly
Present 5-8 price points sequentially using branching logic: start at middle price, branch up on "yes," branch down on "no." Each respondent sees 3-5 questions. Apply Newton-Miller-Smith discount factors: 75% for “definitely buy,” 25% for “probably buy,” 0% for others. Plot adjusted revenue curve (price × adjusted probability × market size). Peak = revenue-maximizing price. [src6] [src3]
Example (1,000 addressable, NMS-adjusted):
| Price | Raw WTP | Adjusted | Est. Customers | Revenue |
|--------|---------|----------|----------------|-----------|
| $19/mo | 82% | 48% | 480 | $9,120 |
| $29/mo | 63% | 36% | 360 | $10,440 ← peak
| $39/mo | 44% | 24% | 240 | $9,360 |
| $49/mo | 28% | 14% | 140 | $6,860 |
Without NMS adjustment, peak appears at $39 — overstating by 34%
Verify: 150+ responses; demand curve downward-sloping; revenue curve has clear peak within VW range · If failed: If everyone says yes, extend range upward 50-100%; if steep dropoff, revisit value proposition [src6]
Duration: 1-2 days · Tool: Google Sheets
Overlay Van Westendorp acceptable range with Gabor-Granger revenue peak. Segment results by customer type (company size, role, industry) to identify tiering opportunities. Compare against competitor pricing. Select 2-3 test prices for live experiment if proceeding to Path C/D. [src4]
Verify: GG peak within VW range; 2-3 test prices selected with rationale · If failed: If VW and GG contradict, prioritize GG (models revenue directly); investigate whether segment mixing is masking two buyer groups [src4]
Duration: 1 week design + 4-6 weeks run · Tool: Stripe, PostHog, Google Sheets
Choose safe experiment structure: new-customer-only test (safest), geographic split, cohort by week, or feature-gated test. Create separate Stripe Price objects per variant. Track conversion rate, ARPU, revenue per visitor, churn. Min 200 conversions per variant. Chi-square for conversion; t-test for RPV. Target p < 0.05. Do not peek and stop early. [src7] [src2]
Safe experiment structures:
├── New-customer-only test (SAFEST)
│ └── New signups get new price; existing keep old
├── Geographic test
│ └── Different prices in different markets
├── Cohort test
│ └── Alternate by signup week
└── Feature-gated test (SAFEST alternative)
└── Same base price, different bundles
Verify: 200+ conversions/variant; 4+ weeks; no audience contamination · If failed: If <100 conversions after 4 weeks, traffic too low — rely on survey results [src2]
Duration: 3-5 days analysis + 1 week implementation · Tool: Sheets, billing system, all channels
Use revenue per visitor as primary metric — captures both price and conversion impact. A 20% price increase with 10% conversion drop is still +8% RPV. Update all channels simultaneously. Grandfather existing customers 6-12 months. Brief sales/CS before pricing page goes live. Monitor 30-60 days post-launch. [src7] [src2]
| Metric | Variant A ($29) | Variant B ($39) | Variant C ($49) |
|---|---|---|---|
| Conversion rate | 32.0% | 25.0% | 15.0% |
| Revenue/visitor | $9.28 | $9.75 | $7.37 |
| Stat sig vs control | — | p=0.04 ✓ | p=0.01 ✓ |
| Est. 12-month LTV | $348 | $468 | $588 |
| 30-day churn | 2.1% | 2.3% | 3.8% |
Verify: New price live across all channels; sales team briefed; existing customers grandfathered; 30-day monitoring active · If failed: If churn spikes >5% above baseline, roll back for existing; keep new price for new signups
{
"output_type": "pricing_experiment_report",
"format": "document + spreadsheet",
"columns": [
{"name": "vw_pmc", "type": "number", "description": "VW Point of Marginal Cheapness — price floor"},
{"name": "vw_ipp", "type": "number", "description": "VW Indifference Price Point"},
{"name": "vw_opp", "type": "number", "description": "VW Optimal Price Point"},
{"name": "vw_pme", "type": "number", "description": "VW Point of Marginal Expensiveness — ceiling"},
{"name": "gg_revenue_peak", "type": "number", "description": "GG revenue-maximizing price (NMS-adjusted)"},
{"name": "gg_demand_curve", "type": "array", "description": "Adjusted purchase probability per price"},
{"name": "recommended_price", "type": "number", "description": "Final recommended price"},
{"name": "experiment_winner", "type": "string", "description": "Winning A/B variant"},
{"name": "revenue_per_visitor", "type": "number", "description": "RPV for winning variant"},
{"name": "segment_prices", "type": "object", "description": "Per-segment prices if tiered"},
{"name": "confidence_level", "type": "string", "description": "p-value for A/B or sample size for survey"}
]
}
| Quality Metric | Minimum Acceptable | Good | Excellent |
|---|---|---|---|
| Van Westendorp responses | 100 | 150 | 200+ |
| Gabor-Granger responses | 150 | 250 | 400+ |
| Survey response rate | > 10% | > 20% | > 30% |
| VW range clarity (PME/PMC ratio) | < 5x | < 3x | < 2x |
| GG revenue curve has clear peak | Yes (any) | Within VW range | Matches VW OPP |
| Live test conversions per variant | 200 | 500 | 1,000+ |
| Statistical significance (if A/B) | p < 0.10 | p < 0.05 | p < 0.01 |
| Revenue per visitor improvement | > 0% | > 5% | > 15% |
| Methods combined | 1 | 2 | 3 (VW + GG + live) |
If below minimum: Re-distribute survey with $10-$25 gift card incentives; extend collection 1-2 weeks. If live test lacks volume, rely on survey results with 30-day monitoring plan. Directional data from 100 responses is better than no data. [src5]
| Error | Likely Cause | Recovery Action |
|---|---|---|
| Survey response rate under 5% | Poor subject line or no incentive | Add $10 gift card; distribute via LinkedIn DMs and community posts; shorten survey |
| VW curves never intersect | Sample too small or respondents confused | Collect 50+ more responses; add example price context; simplify product description |
| GG: everyone says “yes” to all prices | All test prices below perceived value | Extend range upward 50-100%; add 3 higher price points; re-survey fresh respondents |
| GG: everyone says “no” to all prices | Product value not communicated | Improve product description; verify respondent-customer match; check pricing unit |
| VW and GG results contradict | Segment mixing or different respondent pools | Segment both datasets by company size/role — contradiction usually reveals two buyer groups |
| A/B test not reaching significance after 6 weeks | Insufficient traffic or effect < 5% | If effect < 5%, not commercially meaningful — keep current price |
| Post-launch churn spikes > 5% | Price increase too aggressive | Roll back for existing customers; keep new price for new signups; revisit in 90 days |
| Sales team discovers change from prospect | Internal communication failure | Brief team immediately; create FAQ; delay further changes until process fixed |
| Survey panel responses low quality | Respondents rushing for incentive | Add attention check; filter <60s responses; exclude straight-line answers |
| Component | Free ($0) | Lean ($500) | Standard ($2K) | Advanced ($5K+) |
|---|---|---|---|---|
| Survey tool | $0 (Google Forms) | $25/mo (Typeform) | $0 (Conjointly Basic) | $200/mo (Conjointly Pro) |
| Survey panel | $0 (DIY outreach) | $200 (social ads) | $1,000 (Respondent.io) | $3,000 (professional panel) |
| Response incentives | $0 | $150 ($10 × 15) | $500 ($10 × 50) | $1,000 ($10 × 100) |
| Analysis tools | $0 (Sheets) | $0 (Sheets) | $0 (auto-analysis) | $500 (consultant) |
| A/B test implementation | N/A | N/A | $500 (feature flags) | $1,500 (full setup) |
| Monitoring | $0 (PostHog free) | $0 | $0 | $0 |
| Total | $0 | $375-$500 | $2,000 | $5,000-$6,200 |
Customers who compare notes feel cheated. Amazon tested variable pricing in 2000 and faced backlash so severe the CEO personally apologized. “It introduces an element of unfairness to buyers” that causes direct, permanent harm to brand reputation. [src7]
New-customer-only tests are safest. Geographic splits work if markets are isolated. Feature-gated tests (different bundles at same base price) avoid the fairness issue entirely. [src2]
VW identifies a range but not where revenue is maximized within it. In one case, VW suggested $5.75-$7.67 while Gabor-Granger found $13.89 as revenue-maximizing for the same product — nearly 2x the upper boundary. [src4]
Run both surveys. Companies using multiple pricing research methods achieve ~30% higher revenue growth than those relying on a single method. [src4]
Gabor-Granger and direct WTP questions overstate real purchase intent by 10-50%. If 60% of respondents say “probably buy” at $49, actual conversion will be 15-30%. Revenue models on raw data lead to painful misses. [src8]
Discount: 75% for “definitely,” 25% for “probably,” 0% for others. Then validate with even a small live experiment before full rollout. Real behavior is the only reliable pricing signal. [src8]
Use when a startup or SaaS company needs to actually execute pricing research — design the surveys, collect the data, plot the curves, calculate the revenue-maximizing point, run the experiment, and implement the validated price. Not a strategic overview, but step-by-step execution with tools, sample sizes, and statistical requirements. Requires an existing product with users/prospects. Most urgent before fundraising (investors scrutinize pricing power), when unit economics are failing, or when preparing to launch tiered pricing.