.jpg)
August 22, 2025

This week, we hosted a live workshop with Cyrus Nouroozi, founder of Zenbase (YC S24), on the realities of prompt engineering for AI customer support. The session dug into what actually works when designing prompts that scale, from fundamentals to debugging strategies and future directions for the field.
Many teams hit the same wall: solving one prompt issue only to break something else. Support conversations tend to follow a predictable distribution:
The goal is not to automate everything, but to free human agents to focus on high-value, frustrating edge cases.
To move beyond “prompt hell,” Cyrus emphasized the importance of error analysis:
Human-labeled examples then become few-shot training data for evaluation loops.
Prompting is never one-and-done. Key practices include:
The cycle is simple but essential: try, fail, label, adjust.
One surprising insight: XML can outperform Markdown in certain cases where clear section boundaries are critical.
Explicit <tags> create unambiguous boundaries for LLMs, making it easier to group related instructions or constraints.
Cyrus found that XML structure was especially effective for tasks requiring strict sectioning, such as “do/don’t” lists, where the model must clearly recognize where one instruction ends and another begins. In those cases, XML structure provided enough clarity that GPT-4.0 Mini outperformed GPT-4.0 on the same task.
Zenbase has even open-sourced a library to make XML prompting easier, with which you can compose dictionaries, objects, or arrays, then auto-convert them into structured XML.
You cannot improve what you do not measure. Cyrus recommended:
Cyrus sees a clear trajectory:
Better UX is on the horizon: instead of writing prompts, users will provide conversations or goals, and models like Claude or GPT will generate structured prompts automatically.
Customer feedback can serve as live error analysis, but requires careful filtering. Cultural differences mean thumbs up/down signals are not always reliable.
Either way, feedback loops are critical to keeping prompts effective in real-world use.
A big thank you to Cyrus for sharing his expertise, and to everyone who joined us live.