.jpg)
September 3, 2025
This week, we hosted a live workshop with Marcus Storm-Mollard, co-founder and CEO of Clarm (YC X25), on how context turns raw AI capability into reliable, production-grade customer support. We covered why context windows matter, how to deploy trustworthy agents in high-stakes industries, and where deep research architectures are pushing support next.
LLMs are brilliant pattern matchers, but they do not know your customer, product, or policy unless you ground them. Even with large context windows, accuracy degrades if the agent is fed too much, the wrong things, or material untethered to the active user or case. Reliability comes from better context engineering, not just more tokens.
In regulated environments, trust is earned one narrow workflow at a time. Start with a constrained task, such as a single form or policy answer, instrument it for quality, and expand only after sustained accuracy. Zero-hallucination design is primarily scope discipline plus tight retrieval, not an attempt to replace entire roles on day one.
Shipping early matters. The fastest improvements come from putting agents in users’ hands and watching the first two queries obsessively. If those feel smooth and on target, adoption follows. Broad exposure also surfaces unexpected power users, such as developers on private repos, which guides where to deepen product capabilities.
What to focus on:
Beyond cost, AI-native support changes hiring and workflows. As agents absorb repeatable work, product and support loops compress. Startups bias toward builder-generalists and founding engineers who blend engineering, product, and design, reducing handoffs and speeding the path from customer signal to spec to tested flow.
Org shifts to expect:
Simple copilots and chatbots handle FAQs. The hard value comes from deep research agents that plan, retrieve across sources, synthesize, and verify, especially for technical products and regulated domains. These architectures unlock higher-order resolutions, multi-step, multi-doc, account-aware, and the largest efficiency gains in support.
Design notes:
Onboarding pain is often a documentation problem, not just a training problem. Capturing expert workflows, screens, steps, and conversations, and turning them into living docs and teachable traces gives agents and new hires the same reference point, enabling guided, on-device assistance from day one.
If you are worried about risk, start with low-stakes journeys, such as public policies and eligibility rules, and only then bridge to flows involving account data and actions. Reliability is a ladder of trust, not a leap, so earn it through validated small wins.
A phased rollout:
The north star is time saved for customers and teams. AI agents deliver instant, 24x7 responses, while humans focus on relationship-rich work, such as complex escalations, process fixes, and moments that build loyalty. The more mundane tasks AI absorbs, the more human the support experience can feel.
A big thank you to Marcus for the insights, and to everyone who joined us live.