9 Deals. $3.239M Analyzed. The Evidence for What Must Change.
Q4 2025 + Q1 2026 · 5 Lost Deals ($1.925M) + 4 Won Deals ($1.314M) · Prepared for the ELT
How confident are the conclusions? Every claim is tagged by evidence quality.
Tier 1 High Confidence
Recorded calls with Grain AI notes, CRM data, and either verified debrief or active engagement. 4 deals: Global Payments, Vimeo, Fenergo, Check Point.
Tier 2 Directionally Strong
Partial call documentation (typically 2 of 3-8 calls captured), CRM data, no verified debrief. Gaps exist in the middle of each sales cycle. 5 deals: First Advantage, JFrog, Postman, Contentful, WP Engine.
CRM Only Low Confidence
AE self-reported in HubSpot. 4 of 5 loss reasons were wrong or misleading. CRM data is the weakest evidence layer.
| Data Point | Source | Confidence |
|---|---|---|
| Grain call notes | AI-generated summaries from recordings | Medium Not verbatim transcripts. Key quotes may be paraphrased. |
| CRM loss reasons | AE self-reported in HubSpot | Low 4 of 5 were wrong or misleading. |
| ICP revenue data | Public filings + Latka/Sacra estimates | High for public companies. Medium for private. HubSpot data was wrong for 5 of 10. |
| Verified debriefs | Direct conversation with buyer post-loss | High But only First Advantage has one. |
| Won deal financials | SOW values from HubSpot | High |
All 9 deals analyzed, with outcomes and evidence quality.
| Deal | Value | Owner | Outcome | Evidence |
|---|---|---|---|---|
| First Advantage | $500K | Dan Smith | Lost to Value Selling (Val AI) | Tier 2 Verified debrief with Ann + 3 recorded calls + CRM |
| JFrog | $400K | Mike Boogaard | Lost "preferred MEDDIC more central" | Tier 2 2 of 5 calls documented + CRM |
| Postman | $375K | Dan Smith | Lost "requirements changed" | Tier 2 2 of 3 calls documented + CRM. 39-day gap undocumented. |
| Contentful | $400K | Rachael Smith | Lost to Sandler (CRO relationship) | Tier 2 2 of 8 calls documented + CRM. Sandler known 3 weeks before loss. |
| WP Engine | $250K | Dan Smith | Lost "price about triple" | Tier 2 2 calls documented + CRM |
| Global Payments | $414K | Dan Smith | Won multi-SOW expansion | Tier 1 Multiple recorded calls + active engagement |
| Vimeo | $134K | Betsy Peters | Won implementation stalled | Tier 1 Multiple calls + detailed implementation notes |
| Fenergo | $290K | Mike Boogaard | Won land-and-expand | Tier 1 10+ calls documented + SOW progression |
| Check Point | $175K won / $295K lost | Mike Boogaard | Split bought IP, built labor in-house | Tier 1 Won engagement documented. Lost scope is CRM-only. |
Every lost deal surfaced the same structural limitation from a different angle.
| Lost Deal | What They Needed | The Gap |
|---|---|---|
| First Advantage ($500K) T2 | AI agent to certify 551 reps on new product launches at scale | No AI certification product. Val won. |
| JFrog ($400K) T2 | SPICED integrated inside their existing MEDDIC architecture | No MEDDIC integration product. Cosmetic relabeling failed. |
| WP Engine ($250K) T2 | Foundational enablement for 500 reps at ~$200/seat | No scalable digital delivery. Price was 3x competitor. |
| Postman ($375K) T2 | Revenue Architecture for 700-person GTM org, ready for Feb SKO | No mutual action plan. CRO left. 39 days silence. |
| Contentful ($400K) T2 | Skills reinforcement on existing Command of the Message + MEDDIC | No "skills-only" offering. Pitched methodology to a buyer who said "no new methodology." |
| Won Deal | What Worked | Why |
|---|---|---|
| Global Payments ($414K) T1 | SPICED embedded in Salesforce + Gong + Gong Enable. AI scorecards. | Technology integration created switching costs. SPICED is infrastructure, not training. |
| Fenergo ($290K) T1 | ICP analysis, motion validation, revenue risk modeling. | Solved the actual business problem with evidence, not philosophy. |
| Vimeo ($134K) T1 | SPICED/MEDPIC integration. Standalone Sales Ready Messaging. Crawl-walk-run. | Coexisted with existing methodology. Matched the buyer's chaos. |
| Check Point ($175K) T1 | Specialized SKO breakouts with MEDIC integration. Modular, not monolithic. | Bought the IP they couldn't build. Built the generic training in-house. |
WbD is running two fundamentally different selling motions simultaneously. Only one of them wins.
| Element | Evidence |
|---|---|
| Team size | 4 people on Fenergo (Boogaard, Gordillo, Barber, Liem). 4+ on Check Point. T1 |
| Call volume | 10+ calls on Fenergo. 10+ on Check Point. Weekly syncs. T1 |
| Approach | Data-driven diagnostic. ICP analysis, revenue risk modeling, motion validation. Solves the business problem, then proposes methodology. T1 |
| Stakeholder depth | 6 stakeholders at Fenergo. 13+ at First Advantage finalist presentation. T1 |
| Phasing | SOW1 ($120K) earns SOW2 ($170K) at Fenergo. SKO v1 earns SKO v2 + Bangkok at Check Point. T1 |
| MEDDIC handling | At Check Point, MEDIC explicitly positioned alongside bow-tie (Jan 29 call). Integration, not hierarchy. T1 |
| Outcome | Fenergo: $290K won. Check Point: $175K won. |
| Element | Evidence |
|---|---|
| Team size | Solo AE on WP Engine (Dan Smith). Solo on Postman (Dan Smith). Solo + inactive AE on Contentful (Rachael Smith). T2 |
| Call volume | 2 calls on WP Engine (19 days). 2 calls on Postman (39-day gap). 2 documented of 8 on Contentful. T2 |
| Approach | Methodology pitch. Lightboard presentation on SPICED/bow-tie. Comprehensive proposal. T2 |
| Stakeholder depth | 3 stakeholders at WP Engine. 2 at Postman. 7 at Contentful (Dec 5 only). T2 |
| Phasing | $344K single proposal at WP Engine (no entry ramp). $375K single proposal at Postman. $400K CKO + quarterly at Contentful. T2 |
| MEDDIC handling | At JFrog, SPICED positioned above MEDDIC, then hidden behind MEDDIC labels. At Contentful, pitched bow-tie OS to a buyer who said "no new methodology." T2 |
| Outcome | WP Engine: lost. Postman: lost. Contentful: lost. |
Betsy's Vimeo win ($134K) followed neither motion exactly. She read the room (internal chaos, MEDPIC entrenchment, regional misalignment) and adapted: skills-focused delivery, standalone assets (Sales Ready Messaging), crawl-walk-run timeline, architectural SPICED/MEDPIC integration instead of hierarchy. This was improvisation by an experienced deal owner, not an organizational playbook. The Vimeo approach is replicable only if it is productized; currently, it lives in Betsy's judgment.
Three deals tested MEDDIC coexistence. The outcomes diverge sharply.
Tier 1
Why it worked: Genuine architectural specificity. Neither methodology on top. Clear boundaries.
Tier 2
Why it failed: Aug 12 positioning was correct but could not withstand challenge from a skeptic with firsthand evidence of prior WbD failure.
Tier 2
Why it failed: The buyer said "reinforce what we have." WbD said "here is something new."
WbD's consulting model wins above its own ICP and loses within it.
6 of top 10 won deals are above ICP (>$2B)
5 of 5 lost deals are within ICP ($50M-$2B)
| Deal | Revenue | In ICP? |
|---|---|---|
| First Advantage | $1.57B | Yes |
| JFrog | $532M | Yes |
| WP Engine | ~$400M | Yes |
| Postman | ~$313M | Yes |
| Contentful | ~$219M | Yes |
| Deal | Revenue | ICP Status |
|---|---|---|
| Global Payments | $9.3B | Above |
| Teleperformance | $10.2B | Above |
| Hexagon | $5.9B | Above |
| Duracell | $2-3B | Above |
| Check Point | $2.73B | Above |
| Canva | $3.5B | Above |
| Vimeo | ~$417M | In ICP |
| Fenergo | ~$163M | In ICP |
| Infinx | ~$170-200M | In ICP |
WbD defined its ICP as $50M-$2B, but its delivery model is priced and staffed for $2B+ enterprises.
Either the ICP definition is wrong (WbD's real ICP is $2B+), or the delivery model needs a new tier for the $50M-$2B segment. The five lost deals suggest the second interpretation. These were real buyers with real budgets and real problems SPICED solves. The model, not the market, was the mismatch.
4 of 5 CRM loss reasons were wrong or materially misleading.
| Deal | CRM Loss Reason CRM | What Actually Happened Forensic |
|---|---|---|
| First Advantage | "Sandler or Value Selling, incumbent from Sterling" | Sandler was "super outdated" and eliminated early. Value Selling won on Val (AI agent), not incumbency. Pricing was equivalent across all vendors. The CRM framed this as an incumbent legacy fight; it was actually an AI product competition. |
| JFrog | "Preferred MEDDIC more central to training" | A surface-level attribution hiding 4 deeper factors: (1) prior failed WbD implementation with a key stakeholder, (2) JFrog already building MEDDIC into Salesforce before WbD entered, (3) WbD proposed methodology when JFrog needed forecast accuracy, (4) cosmetic relabeling signaled desperation. |
| Postman | "Requirements changed" | Requirements did not change. The CRO went on paternity leave. No bridge plan existed. 39 days of silence followed. No evidence of any scope or strategic change. This is a polite CRM euphemism for "the deal died of neglect." |
| WP Engine | "Price about triple similar competitor" | Accurate on the surface but incomplete. Pipeline was $250K; proposal was $344K (38% over WbD's own estimate). No phased option offered. No scope negotiation occurred. The deal died in 3 days with no counter. The price was wrong because the packaging was wrong. |
| Contentful | "CRO had longstanding Sandler relationship, shaped the brief" | Most accurate of the five, but incomplete. Rachael knew about Sandler 3 weeks before the loss (Katie told her Nov 19). No counter-strategy was executed. WbD pitched a bow-tie OS to a buyer who explicitly said "no new methodology." The relationship was deterministic, but the positioning mismatch made it easy for the CRO to justify Sandler on substance, not just loyalty. |
If the ELT reads the CRM at face value:
Five different problems suggesting five different fixes.
What the forensic evidence actually shows:
One systemic problem manifesting five ways. WbD has a single delivery model (high-touch consulting) that cannot flex for scale buyers (WP Engine), methodology-entrenched buyers (JFrog, Contentful), AI-demanding buyers (First Advantage), or time-constrained buyers (Postman). The CRM obscures this pattern because each AE picked a different dropdown reason for the same structural failure.
The evidence from all 9 deals converges on a single product thesis.
Value Selling's AI agent Val was "the deciding factor" in a $500K deal. Val could "certify entire GTM org on new product launches at scale." WbD positioned AI as a feature within training. Value Selling positioned AI as the product. For a 551-person post-M&A org, the AI product won.
WbD's largest customer is assembling the SPICED AI product by hand: Gong Enable uses real call recordings to create AI training environments. SPICED scorecards evaluate reps against methodology criteria automatically. Enable creates certification paths from actual customer conversations. This is bespoke systems integration costing $414K. A productized version would be deployable at any Gong customer in weeks, not months.
WP Engine asked for SPICED by name. 500 CS reps. Entry-level. Needed foundational methodology at scale. WbD's only option: $344K comprehensive consulting for 25 people. At $200/seat for 500 reps, SPICED AI captures $100K, within WP Engine's budget, competitive with the alternative, and scalable to the full org.
A customer built SPICED-powered analytics using Claude (Anthropic's AI) without WbD's direct involvement. The methodology's intellectual property is accessible enough that technically sophisticated buyers can reconstruct it in AI tools. This is simultaneously a validation of the methodology's power and a warning that the IP is not defensible without a product wrapper.
Check Point paid $175K for the specialized, methodology-integrated SKO content and $40K for an IP license. They built the other $295K of generic training in-house, despite their own data showing "10 training sessions, 68 hours, zero performance change." Buyers want the IP. They will build the delivery themselves. A SPICED AI product is the IP delivered at scale without requiring human consulting labor.
Contentful's CRO "recently created a call structure framework using ChatGPT." Executives are already using AI to build methodology tools. They will either build them on SPICED (if WbD provides the platform) or on generic principles (if WbD does not).
| Lost Deal | Reps | At $200/seat | vs. Lost Revenue | Captures? |
|---|---|---|---|---|
| First Advantage | 551 | $110K | vs. $500K lost | PARTIAL. Competes with Val, may not win alone. Needs AI certification feature. |
| JFrog | 160 | $32K | vs. $400K lost | PARTIAL. Entry point. Needs MEDDIC-native integration. Upsell to consulting. |
| WP Engine | 500 | $100K | vs. $250K lost | YES. Exactly their budget range. Foundational methodology at scale. |
| Postman | 700 | $140K | vs. $375K lost | PARTIAL. Entry point while CRO is away. Self-serve. Bridge the gap. |
| Contentful | ~200 (est.) | $40K | vs. $400K lost | PARTIAL. Skills reinforcement without methodology branding. "Boil the frog." |
Conservative estimate: $200/seat SPICED AI captures $100K-$422K across the 5 lost deals depending on feature depth. It does not replace the consulting model for Fenergo/GPN-style engagements. It creates a new market tier that does not currently exist.
Three tiers of action, ordered by strategic impact.
A technology-delivered SPICED experience that lives inside existing tools (Gong, Salesforce, LMS platforms). Val already exists at Value Selling. GPN is building it manually. The market will not wait.
What must change in how WbD sells, based on the evidence.
CRM data is actively misleading leadership. 4 of 5 loss reasons were wrong.
The most powerful verbatim statements across all 9 deals. Grain quotes are AI summaries, not raw transcripts. Exact wording may be paraphrased, but substance is verified.
"Val was the deciding factor."
Ann, First Advantage (verified debrief) Tier 1
The single most important data point in the entire analysis. An AI agent decided a $500K deal.
"Val could certify entire GTM org on new product launches at scale."
Ann, First Advantage (verified debrief) Tier 1
This is what WbD cannot do today and what SPICED AI must do tomorrow.
"Confusing due to WbD's inexperience with MEDDIC."
Christian Arrigo, JFrog (Call 5) Tier 2
A prior customer's verdict on WbD's MEDDIC integration capability. This is not a positioning problem. It is a product problem.
"NOT looking to introduce new methodology. Goal is to reinforce existing framework."
Katie Van Hoomissen, Contentful (Call 1) Tier 2
The buyer told WbD exactly what they wanted on the first call. WbD pitched something else.
"Price about triple similar competitor."
CRM close note, WP Engine CRM
A company that asked for SPICED by name walked away in 3 days because the only way to buy it was a $344K consulting engagement.
"Sales skills are relatively the same."
Chris Masino, CRO, Contentful (Call 5) Tier 2
The CRO wanted one consistent program. WbD proposed three differentiated tracks. Sandler offered one.
"Dive and catch culture, siloed org, lack of defined processes."
Doug Lavanchy, Postman (Call 3) Tier 2
A Revenue Architecture buyer describing themselves in Revenue Architecture language. The need was real. The execution was not.
"10 training sessions, 68 hours, zero performance change."
Check Point internal data Tier 1
Their own generic training failed. They still built more of it in-house ($295K). But they bought WbD's specialized IP ($175K). Buyers build labor in-house. They buy intellectual property from vendors.
"Entry-level reps lack foundational sales training. First-time managers: homegrown, learned on the job."
WP Engine (Call 1) Tier 2
The most foundational buyer need possible, met with the most premium delivery model possible. This is the product-market mismatch in one quote.
"Hard decision between WbD and Value Selling."
Ann, First Advantage (verified debrief) Tier 1
WbD was close. Not a distant third. One capability gap (AI certification at scale) separated $500K won from $500K lost.
Click any deal to expand its full forensic profile.
What they came for: Wanted an AI agent to certify 551 reps on new product launches at scale in a post-M&A integration.
What they needed: AI-powered certification at scale. Val could do it. WbD could not.
Verdict: WbD offered "93 AI partners certified on SPICED" and AI role-play scenarios, positioning AI as a feature within training. Value Selling positioned AI as the product. Val was "the deciding factor." Pricing was equivalent across all vendors. This was an AI product competition, not a legacy incumbent fight.
What they came for: SPICED integrated inside their existing MEDDIC architecture for a PLG-to-enterprise growth mandate.
What they needed: Forecast accuracy and MEDDIC-native integration, not methodology hierarchy.
Verdict: WbD's positioning evolved from "SPICED is the operating system with MEDDIC fitting inside" (hierarchy claim) to the correct framing ("SPICED enables MEDDIC criteria") to retreat under pressure ("hide SPICED behind MEDDIC labels"). Christian Arrigo's prior negative WbD experience was not discovered until Call 5. The Aug 12 positioning was correct but could not withstand a skeptic with firsthand evidence of prior failure.
What they came for: Revenue Architecture for a 700-person GTM org, ready for a February SKO.
What they needed: A mutual action plan with dates and owners. The CRO went on paternity leave. No bridge plan existed.
Verdict: CRM says "requirements changed." Requirements did not change. 39 days of silence followed the CRO's departure. No evidence of any scope or strategic change. Doug Lavanchy described the exact problem WbD solves ("dive and catch culture, siloed org, lack of defined processes"). The deal died of neglect, not changed requirements.
What they came for: Skills reinforcement on existing Command of the Message + MEDDIC. Explicitly stated "NOT looking to introduce new methodology."
What they needed: A "skills-only" offering that reinforced their existing framework without new branding.
Verdict: WbD pitched the bow-tie operating system with 3 role-based tracks to a buyer who said "no new methodology." The CRO had a longstanding Sandler relationship (known 3 weeks before loss). No counter-strategy was executed. The CRO wanted uniform skills training ("sales skills are relatively the same"). Sandler's consistent-methodology model aligned with his preference. WbD's differentiated proposal aligned with a sales leader (Brett McNay), not the CRO who controlled budget.
What they came for: Foundational enablement for 500 CS reps at ~$200/seat. Asked for SPICED by name.
What they needed: Scalable digital delivery. Entry-level reps lacking foundational sales training. First-time managers who were "homegrown, learned on the job."
Verdict: Pipeline was $250K. Proposal was $344K (38% over WbD's own estimate, $13,760/seat for 25 people). No phased option offered. No scope negotiation occurred. The deal died in 3 days with no counter. The price was wrong because the packaging was wrong. At $200/seat for 500 reps, SPICED AI would have captured $100K, exactly within their budget range.
What they came for: SPICED embedded in their technology stack for ongoing revenue operations.
What they needed: Technology integration: SPICED in Salesforce + Gong + Gong Enable. AI scorecards evaluating reps against methodology criteria.
Verdict: Technology integration created switching costs. SPICED became infrastructure, not training. This is the prototype for what a SPICED AI product looks like: scorecards in Gong, AI certification scenarios in Enable, methodology fields in Salesforce. GPN is paying $414K for WbD to build it manually, one customer at a time.
What they came for: Help with internal chaos, MEDPIC entrenchment, regional misalignment.
What they needed: SPICED/MEDPIC architectural integration. Standalone Sales Ready Messaging asset. Crawl-walk-run phasing.
Verdict: Betsy read the room and adapted. Coexisted with existing methodology. Did not require methodology adoption. Delivered skills-focused content (question cards, pocket stories by persona) that embeds SPICED thinking without requiring anyone to learn SPICED as a brand. This is the "boil the frog" approach that works.
What they came for: Help with 40% churn, growth concentration, and a 322-day sales cycle.
What they needed: Data-driven diagnostic: ICP analysis, motion validation, revenue risk modeling.
Verdict: Solved the actual business problem with evidence, not philosophy. SOW1 ($120K) earned SOW2 ($170K) through demonstrated results. 6 stakeholders engaged. Land-and-expand model worked exactly as designed. This is the Boogaard Motion at its best: multi-person team, diagnostic depth, phased commitment.
What they came for: SKO content and ongoing enablement for their global sales org.
What they needed: Specialized, methodology-integrated SKO breakouts. MEDIC positioned alongside bow-tie. Modular, not monolithic.
Verdict: Check Point bought the IP they couldn't build ($175K + $40K IP license) and built the generic training in-house ($295K), despite their own data showing "10 training sessions, 68 hours, zero performance change." This confirms the thesis: buyers want intellectual property from vendors and build delivery labor themselves. A SPICED AI product is the IP delivered at scale.
Win/Loss Forensic Synthesis · April 2, 2026 · Prepared for the WbD Executive Leadership Team
Sources: Grain call recordings, HubSpot CRM, verified debriefs, public revenue data. Evidence gaps noted where they exist.