A Thesis · For Clay · By Sidharth Sundaram

The smartest dumb pipe in B2B GTM. Your enrichment layer is more capable than you're treating it.

A twelve-week internship thesis for the Enrichment & AI PM role. Clay's enrichment layer runs more than a billion requests through 75+ providers. None of that signal compounds. Here's the architecture that makes it learn — without new infrastructure and without taking control out of Ops's hands.

Twenty minutes to walk through it →
§ I · The diagnosis

Stateless by default.

Clay's enrichment layer is the most data-rich pipe in B2B GTM. Seventy-five-plus providers. More than a billion requests. Every domain pattern, every industry, every account stage. None of it aggregates. The waterfall doesn't learn. The marketplace can't tell you which providers win for which segments. Claygent doesn't know which research patterns yielded replies last quarter. Each request runs as if it were the first.

Exhibit 01 · from my own Clay workspace

Eight providers. One wins everything that wins.

I ran a real Find-People table on doorstep.aiin my own Clay account. The Function cascades through eight email-finder providers: LeadMagic, Findymail, Prospeo, DropContact, Hunter, Datagma, Icypeas, SMARTe. Findymail at position 2 won every row that produced a valid email (5 of 8). LeadMagic at position 1 failed all 8. Positions 3–8 contributed zero. The next time this Function runs on a .ai domain, the cascade restarts at LeadMagic. The system never learned.

find_people· doorstep.ai · 8/8 rowsemail cascade · 8 providers
Person
LeadMagic
1
Findymail
2
Prospeo
3
DropContact
4
Hunter
5
Datagma
6
Icypeas
7
SMARTe
8
Daley Ervin
Shashwat M.
Hrach Simonian
Pete Martin
Yuri Choi
Sheel Patel
Jerin Mathew
David Ulloa
wins0/85/80/80/80/80/80/80/8
Findymail at position 2 wins everything that wins. LeadMagic at position 1 fails all 8. Positions 3–8 contribute zero. 34 billed lookups returned 5valid emails — 15% efficiency. The next .ai domain runs the same cascade.

Look closely at the column headers in any Clay table: 100% / 88% / 63% / 89%. Clay already tracks freshness per provider column — per-provider telemetry stored at the column level. Different providers produce data with different staleness curves; that's a quality signal in itself. Clay doesn't aggregate it across customers, and it doesn't segment it by industry or stage. The columnar primitive is shaped to be joinable. Marketplace intelligence is a join, not a build.

Captured 2026-04-25 · verifiable in any Clay workspace
Exhibit 02
Viewing a function in 'Live' mode mainly serves as an 'audit log' of data processing. Once run, rows are effectively archived and cannot be modified from within the function itself.
Clay University, Functions documentation
university.clay.com/docs/functions

Live mode is an audit log, not an analytics surface. Function-level performance data has no home in Clay’s schema today.

Exhibit 03
When we make updates to add additional context or refine prompts, the optimizations are automatically rolled out to all of the workflows using the function.
Osman Sheikhnureldin · Head of GTM Ops, Clay
clay.com/blog/functions

“We” means humans editing. Refinement is Ops typing into the editor. There is no automated signal informing those edits today.

The negative space

Across every Clay University Functions and integration lesson, these enrichment-quality words return zero hits each.

win ratesuccess rateoutcomeperformancelearningprioritizerefineiterate

The forward-looking lesson “What's Possible Next: Functions Beyond the Table” describes more distribution surfaces. It does not describe learning.

§ I.5 · Drafts that died

The first version of this thesis was about MCP.

It centered on a single phrase from the April 22 MCP launch post — “creating a tighter feedback loop between Ops and sales teams”— and read it as a promise Clay shipped in only one direction. The argument was clean. The artifact ran on the rhetoric of quoting Clay back to themselves.

It died for two reasons. The first: MCP and Functions sit closer to the rep distribution layer than to enrichment. The role this thesis is for owns Claygent, the marketplace, integrations, and tables. A brilliant pitch in another PM's territory is a polite forward, not an interview.

The second: a single launch-post phrase is parry-able in ten seconds. The structural evidence — the waterfall above, Live mode as audit log, the negative-space audit — is durable. The linguistic teardown isn't. The enrichment-side version of this thesis is stronger because the evidence is in scope, the bet is in scope, and the strongest single piece of evidence is something only an Enrichment PM can act on.

§ II · The prototype

Two views of the same architecture.

The Dashboard is where the intelligence shows up — Ops sees recommendations joined from cell-level data across thousands of customer tables. The Builder beneath shows what gets captured at each cell to make the join possible. The first recommendation in the Dashboard (“saves ~16 credits per row on .ai Series A–B accounts”) is what marketplace intelligence would have produced running on the waterfall in §I. Read top-to-bottom.

⚡ FunctionsMarketplace intelligencenew tab · proposedSidharth's Workspace · last refreshed 4 min ago
47% of your email-finder spend is going to providers that don't return value for your accounts. Three pending recommendations could close the gap by ~$1,200 / mo.
Credit efficiency · 30d
53%
↑ from 41%
Wasted credits · 30d
4,200
≈ $420
Pending recs · est. savings
$1.2k/mo
3 to review
Applied this month
7
+11% hit-rate
Pending recommendations
All require Ops approval before applying
Saves ~16 credits per row on .ai Series A–B accounts
Findymail returned a verified email on 73% of 142 runs across .ai Series A–B accounts last quarter. Hunter, currently in position 1, won 31%. Promoting Findymail to position 1 cuts the cascade short on the segment where it wins.
Est. +14% waterfall hit-rate · ~$320 / mo at current volumen=142 · last 90 days
Removes 47 stale rows / month from biotech sequences
BuiltWith returned tech-stack data older than 180 days on 23% of runs against biotech > 200 HC. Demoting it past cleaner sources cuts stale outputs without sacrificing coverage.
Est. -47 stale rows / mo · downstream sequence quality upn=312 · last 60 days
+18% reply rate at the same credit cost
Hiring-signal prompt v2 wins replies 31% more than v1 on Series B+ accounts. Same Claygent run, same cost — different prompt scaffold. v3 currently in test, +11% over v2 at low n.
Est. +18% reply rate · zero credit deltav1 n=142 · v2 n=89
Provider win rate · email finders · last 90 days
% verified return
SegmentHunterFindymailApolloLeadMagicBuiltWith
AI Infra · Series A–B31%73%44%12%
Devtools · Series B+58%62%71%39%44%
Biotech · 200+ HC12%8%29%55%23%
FinServ · Enterprise67%54%82%61%71%
Healthtech · Growth41%38%33%47%62%
What changed this month
Apr 24
Findymail promoted on .ai Series A–B
Sara approved · +14% observed hit-rate this week
applied
Apr 18
Apollo demoted on biotech > 200 HC
Sara approved · sequence quality up 9%
applied
Apr 11
Hiring-signal v2 promoted to default for Series B+
Sara approved · +18% reply rate observed
applied
Apr 03
ZoomInfo demoted on early-stage AI
Sara dismissed · prefers manual control here
dismissed
Recommendations driven by 1.2M aggregated enrichment runs across 75+ marketplace providers✦ Ops-controlled · nothing applied without your click
Fig. 1 · Marketplace intelligence
A new tab inside Functions — additive, not overriding existing Function authoring. Provider performance by segment, prompt-archetype comparisons, applied-vs-dismissed timeline. Every recommendation requires Ops approval; nothing self-applies. The first recommendation here ('saves ~16 credits per row on .ai Series A–B') is what marketplace intelligence would output running on the waterfall in §I.
§ III · The hypothesis

Marketplace intelligence. Provider performance by segment.

Capture the disposition signal Clay's pipes already produce. Aggregate provider win rates by industry, geography, account stage. Aggregate Claygent prompt-archetype performance by Function. Surface both as recommendations to Ops — not auto-applied changes. “Findymail won 73% on Series A–B .ai domains last quarter; Hunter won 31%. Promote Findymail to position 1 in your waterfall?”Ops decides. Clay provides intelligence Ops didn't have before.

This is a competitive asset no one else can replicate. Seventy-five-plus providers in the marketplace, billions of requests across every domain pattern and industry, and the implicit quality signal of which provider returned what for whom. No competitor has Clay's query volume; no competitor can replicate usage-informed routing without first becoming Clay. The advantage compounds with every enrichment run.

And no, this is not a competing surface to Sculptor. Sculptor today is stateless — every “why is this row here?” regenerates fresh, with no memory of past outcomes. Marketplace intelligence is the substrate. Sculptor stays the consumption surface, finally with something to remember.

Fig. 3 · the loop
FunctionsOutputsCRMMarketplaceCLOSE THE LOOPeach cycle compounds the next
exists todaythe missing arcs

Functions produce Outputs. Outputs flow into CRM as opportunities, replies, closed-won. Today the cycle stops there. Closing the dashed arc turns the marketplace into a node that accumulates — signal fills it as customers run Functions, recommendations refine routing, refined routing produces cleaner outputs, cleaner outputs become better disposition data. Each cycle makes the next cycle better. The dynamic that eventually counterweights it — what happens when everyone routes to the same provider — is named in §III.5.

How it actually ships

One governing principle runs through everything below: Ops stays in control. The system surfaces signal. Ops decides whether to apply it. Nothing in this proposal silently rewrites a Function someone built.

  1. i.
    Reasoning capture is a schema change, not a UI change.

    Claygent already computes the reasoning trail every run. Clay's University documentation confirms this: “users can optionally add confidence or reasoning columns” (Clay U) — meaning the trail is already computed; today it's discarded the moment the cell renders unless the user pays for a second column to retain it. Capturing it once means retaining what's already in memory — no new fetches, no new credits. The default table view is unchanged. Reasoning surfaces only when explicitly requested: an opt-in inline panel, MCP response payloads, or CRM exports.

  2. ii.
    Disposition comes from CRM read-back, not rep input.

    Clay's Scheduled Sources already poll Salesforce and HubSpot — “existing rows will be updated with any new information” (Clay U). Toggle that on, re-run a SOQL query against Opportunity stage, and stage transitions surface in the table. The infrastructure exists. The wiring doesn't. Sequencer-side signals (Outreach reply rates, Salesloft meetings booked) aren't readable today — that gap is itself second-order evidence the loop has been outbound-only, and it's v2 territory.

  3. iii.
    Refinement is metadata, not re-enrichment.

    Promoting Findymail to position 1 in a waterfall is a metadata change, applied lazily at the next run. Existing rows are not re-enriched. Following from theses i and ii, outcome capture itself is free — Claygent reasoning is already computed (i); CRM read-back uses primitives Clay already ships (ii). Zero new credit cost. Zero new infrastructure.

  4. iv.
    New surfaces are additive, not overriding.

    Marketplace intelligence renders as a new tab inside the Functions surface, not a redesign of Function authoring. The reasoning panel is opt-in click-to-expand. MCP gets richer payloads — same shape, more metadata. Nothing currently in Clay gets touched.

§ III.5 · Assumptions & risks

Who this is for, and what could go wrong.

Who this is for.

The GTM Engineer or RevOps lead who builds Functions and configures waterfalls. They live inside Clay all day. They feel every dollar of wasted credit when finance reviews the bill. Marketplace intelligence is for them. Reps consume output via MCP or CRM; they don't see provider attribution, and they don't want to. This is explicitly not built for reps.

The pains, ordered.

  1. i.
    Burned credits on dead-weight providers. The waterfall in §I produced 34 billed lookups for 5 valid emails — 15% efficiency.
  2. ii.
    No data to defend provider choices when finance asks why the bill is what it is.
  3. iii.
    Manual waterfall maintenance — provider quality drifts; Ops finds out from anecdotes, not signal.
  4. iv.
    Provider sprawl. 75+ marketplace options. New customers don't know what to enable; existing customers settled into defaults years ago.
  5. v.
    One waterfall, all segments. Biotech and fintech run the same cascade despite dramatically different provider performance per segment.

Why Clay should care.

The retention argument is strongest. “Clay saved us 30% on enrichment credits” is a renewal-securing line. It addresses the credit-cost narrative Clay's own March pricing memo names openly. Differentiation follows: no other orchestrator has cross-customer marketplace data at this scale, and it compounds with every enrichment run.

The subtle revenue risk is real. Helping customers spend less per row could shrink per-customer credit revenue. The honest counter: marketplace intelligence is the kind of feature that moves a customer from using Clay for one workflow to trusting it across their entire GTM stack. The lifetime-value gain from that bucket transition materially exceeds the per-row credit savings. Power users don't churn at credit prices. Casual users do.

What could go wrong.

Partner-ranking tension.
Hunter, Findymail, Apollo and the rest won't love being publicly ranked. Commercial relationships need careful handling — providers should see their own performance privately before any of it is user-facing.
Privacy and federated learning.
Cross-customer aggregation needs anonymization and a TOS review. Some customers will object to their outcomes informing others' recommendations.
Cold start.
Per-account recommendations need critical mass before they're meaningful. The fix is the standard collaborative-filtering pattern: seed with Clay's anonymized cross-customer aggregate, then personalize as per-account volume grows. Solved territory, not novel research.
Statistical significance.
Long-tail segments may never get enough signal — Series C+ logistics in EMEA, e.g. A confidence-threshold mechanic must hold back recommendations until n is meaningful.
Provider arbitrage.
If everyone routes optimally, the optimal providers gain pricing power. The marketplace's job is to keep that competition healthy, not to declare permanent winners.
§ IV · Twelve weeks

What I'd ship.

  1. Wks 1–3
    Diagnostic
    Partner with the enrichment + data eng team to query aggregate waterfall performance: provider win rate by industry, by account stage, by Function archetype. Pair this with 6 customer interviews (Ops champions running heavy waterfall Functions). Output: the actual shape of the loss inside Clay's pipe today.
  2. Wks 4–5
    Architecture
    Marketplace-intelligence schema proposal aligned with eng. The smallest schema change to retain disposition signal from Scheduled Sources. Recommendation surface inside Functions, not a new app. Cost-neutrality and Ops-control reviews.
  3. Wks 6–9
    Ship
    Marketplace intelligence v1 for one enrichment archetype (waterfall email finders) as an A/B test against control accounts. Reasoning capture for one Claygent column type, opt-in surface only.
  4. Wks 10–12
    Validate
    Measure waterfall hit-rate delta on accounts running the recommendation surface vs. control. Propose Function-level recommendation expansion + sequencer-side signal-ingestion roadmap for v2.

Week one deliverable: a joint diagnostic with engineering and the Ops product PM — at which boundary inside the Functions execution path can disposition signal be retained, and what's the smallest schema change to do it?

§ V · Colophon

Sidharth Sundaram.

Master's in Engineering Management at Purdue. Four years B2B PM at upGrad Campus and Interview Kickstart, both at the intersection of AI, learner activation, and product workflows for non-technical end users.

Why Clay: one of the few B2B SaaS bets where creativity is on the org chart, where pricing memos get published, and where the product itself reads as opinionated. The PM Enrichment & AI role sits exactly where I want to live for the next decade.