Field Notes / AdTech
№ 014 · 2026
Essay · Conversational AI · Rebuilding the Ad Stack

From Clicks to Conversations.

AI chat — not CTV — is the real frontier of digital advertising. This essay lays out what changes when intent is expressed rather than inferred, and what has to be rebuilt across the stack: monetization-aware prompt engineering, edge-side context compression, live creative, a split-agent architecture that protects the user experience, and a new set of metrics where CTR is no longer the unit of analysis.

1,496 words7 minApril 24, 2026Vasco Reid, Gad Benram

For the past five years, the digital advertising industry has convinced itself that the future is television. Connected TV budgets have ballooned, every DSP has a CTV story, and publishers have chased walled-garden streaming inventory as if the big screen were the last undiscovered country. It isn't. While the industry has been perfecting 30-second spots for streaming apps, the actual frontier — where user attention, intent, and data are being reforged — has moved somewhere else entirely. It has moved into the chat window.

The shift from static formats to conversational surfaces isn't a feature release. It's a platform-level event, on par with the move from desktop to mobile, and it demands that we rebuild the entire ad stack from the prompt up.

01 / The Old Internet Links. Clicks. Inference. click → guess 02 / The New Internet Dialogue. Disclosure. Intent. intent vector → 0.84 · 0.21 · 0.67 · 0.14 · 0.93 · 0.52
Fig. 01 The old stack inferred intent from breadcrumbs. The new one receives it, in sentences, from the user.

The Death of the Click, the Birth of Conversational Intent

The old internet was a geometry of links. A user clicked, landed, bounced, or converted, and the ad stack's job was to guess — from keywords, cookies, and behavioral breadcrumbs — what that user might have wanted. Intent was always inferred. It was a probabilistic shadow cast by behavior, never the thing itself.

Conversational AI collapses that distance. When a user types a question into a chat interface, they aren't leaving breadcrumbs — they are telling the system, in natural language, exactly what they want, how they feel about it, and what adjacent things they're curious about. "Help me plan a trip to Japan for two weeks in October, mid-budget, my partner doesn't like cities" is not a keyword. It is a full intent dossier, voluntarily disclosed, more valuable than a year of cookie tracking.

This is the single most important economic fact about AI chat: it converts latent intent into expressed intent, and expressed intent is the most valuable substrate advertising has ever had access to.

The Publisher, Reinvented

Most content publishers are still running an architecture conceived in 2005: write an article, surround it with banners, hope for clicks. Traffic is a river they skim from, not a relationship they cultivate. Generative AI offers a way out.

A publisher that embeds a conversational layer into its content turns a static window display into a dialogue. Instead of scrolling past a personal finance article, the reader asks it questions — about their own situation, their risk tolerance, what products might fit. Every exchange is a disclosure. Every disclosure is signal. The publisher stops being a page and becomes a surface where intent is actively revealed rather than passively guessed at.

This is where zero-party data — information the user volunteers — becomes the category's new gold. Unlike cookies (deprecated), pixels (leaky), or fingerprinting (contested), zero-party data is both the highest-quality signal and the most regulation-proof. The user gave it to you. On purpose.

Architecting the New Stack

Recognizing the opportunity is the easy part. Building for it is not. A conversational ad stack is a different animal from a display or video stack, and it needs four distinct capabilities that don't exist off the shelf.

Monetization-aware prompt engineering. The chat agent cannot be a pure assistant. It has to be designed to ask the kinds of follow-up questions that, while genuinely useful to the user, also surface fields that live campaigns care about. "Would you like to learn about alternative investments?" is useful to the reader and actionable for the bidder. Writing that question well — so it serves the user first and the advertiser second — is now a core AdTech discipline, not a UX afterthought.

Edge-side context compression. You can't send an entire conversation transcript to an ad server in real time. The solution is vectorization: compact embedding models running close to the user that turn the ongoing conversation into a dense vector representation. The ad server then performs cosine-similarity matching against campaign intent vectors in milliseconds. It's semantic bidding, not keyword bidding.

Live creative. The 300×250 JPG is dead in this environment. Advertisers stop uploading finished creative and start uploading goals: ICP definitions, product specs, tone constraints, offer parameters. The LLM assembles the ad in-context, in real time, woven into the conversation with wit and relevance. Creative becomes a function, not a file.

Multi-modal ad units. And that creative is no longer bounded by text. Interactive buttons, embedded purchase widgets, visual comparison cards, mini-quizzes — all generated inline, all native to the conversational surface. The ad unit is whatever the moment calls for.

Fig. 02 — The Conversational Ad Stack User expresses intent Conversational Surface Service Agent helps the user — answers questions — retrieves, reasons, resolves Monetization Agent handles the bid — matches campaigns — assembles live creative separation of concerns answer returned Anonymization Layer strips PII before vectorization Embedding Space · Cosine Similarity user intent campaign θ campaign vectors c₁ · 0.81 · 0.24 · 0.70 · 0.18 ... c₂ · 0.42 · 0.88 · 0.15 · 0.61 ... c₃ · 0.79 · 0.22 · 0.68 · 0.22 ... ← match c₄ · 0.11 · 0.54 · 0.92 · 0.38 ... match 0.97
Fig. 02 The pipeline: conversation flows through a split-agent architecture, gets stripped of PII, vectorized, and matched to campaigns by semantic similarity — then assembled back into the surface as live creative.

Technological Independence

There is an uncomfortable truth embedded in all of this: if your entire business runs on OpenAI's or Google's models, you are building an advertising empire on top of companies that have their own advertising empires to build. Their incentives are not yours.

A serious conversational AdTech player cannot be a downstream tenant on a hyperscaler's roadmap. That means either training proprietary models or, more realistically, intelligent fine-tuning on top of strong open-weight foundations — Llama, Mistral, Qwen, whatever the frontier of open models looks like this quarter. Infrastructure sovereignty becomes a strategic moat.

Architecturally, the right pattern is multi-agent. One agent — the service agent — exists to help the user. A second agent — the monetization agent — handles bidding, creative assembly, and campaign matching. Keeping these separated is what preserves answer quality. When a single model is asked to both serve the user and sell to the user, it will eventually compromise the first for the second, and users will feel it. Separating the concerns protects the experience and, paradoxically, protects the advertiser's investment too: a degraded chatbot is a dead channel.

New Economics, New Metrics

The old ad stack ran on fractions of a cent. Serving a banner costs effectively nothing. Serving a conversational response costs real money — every turn is an inference call, and inference is priced in tokens. The entire economic equation has to be renegotiated.

The core equation is unforgiving: expected revenue per conversation must exceed inference cost per conversation by a margin that supports the rest of the business. Get this wrong and unit economics collapse. Get it right and you have a model with dramatically better yield than traditional display, because each conversation contains multiple monetization opportunities stacked across turns.

The metrics follow. CTR becomes almost meaningless. The measurements that matter in this world are different animals entirely.

Fig. 03 — The Metric Shift Yesterday CTR 0.08% click-through rate a clicked link.a guessed want. Tomorrow CPCo $0.42 cost perconversation Intent Lift +24% sentiment shift,in-session Dynamic CPA $11.80 converted in-chat,no landing a sustained dialogue.a disclosed intent.a funnel collapsed into one surface.
Fig. 03 CTR measured an attention flinch. The new triad measures a relationship — its cost, its movement, and its outcome — inside a single conversation.

And a welcome consequence: the Made-For-Advertising (MFA) playbook dies here. MFA sites work because shallow content is cheap and ads are blind. In a RAG-powered conversational surface, the retrieval quality of the underlying content directly determines whether the user stays, re-engages, and converts. Garbage content produces garbage intent, which produces zero revenue. Quality is no longer a virtue — it's a prerequisite.

Privacy, Ethics, and the Regulatory Surface

Handing a model a stream of personal disclosures and then routing that stream to ad servers is a regulatory minefield walked blindly. The architecture has to handle it from day one.

The right pattern is an anonymization layer that sits between the conversational agent and the ad server. PII — names, precise locations, account identifiers — is stripped before the conversation is vectorized. The ad system sees intent, not identity. This is not a nice-to-have; it's the difference between a sustainable business and a GDPR headline.

On the content side, Sponsored RAG — where paid information is integrated into retrieval-augmented responses as genuine, cited knowledge — is the native ad format of this era. Done well, it's more useful to the user than generic content. Done poorly, it's fraud.

Which is why transparency bumpers — clear [Sponsored] markers, provenance citations, and FTC-compliant disclosures — are not optional. Users will tolerate commercial dialogue. They will not tolerate the feeling of having been conned by a machine that pretended to be a friend. The regulatory line here is not going to be drawn by legislators first; it's going to be drawn by the first major trust collapse. Building on the right side of it now is the cheapest option any of us will ever get.


The Old Internet Was Links. The New Internet Is Dialogue.

Most of this is a lot to build. Most of it doesn't have mature vendors. Some of it requires rethinking pricing, creative, measurement, and infrastructure simultaneously. That's what a platform shift looks like from the inside.

What remains constant is the economic engine. Advertising paid for the web, paid for mobile, paid for social, paid for video — and it will pay for this too. The money will not vanish; it will migrate. It will migrate to whoever builds the stack that treats conversation as first-class inventory, that respects the user enough to keep the service agent and the monetization agent separate, and that recognizes zero-party intent as the most valuable signal the industry has ever had.

CTV was never the frontier. It was the last well-lit room in the old house. The frontier is the chat window — and the stack to serve it has barely been written.

End.   Set in Fraunces & Newsreader.
№ 014 · 2026 · AdTech Lab