
Conversational UX vs Agent Experience
A practical guide to the difference between conversational UX and agent experience, and why that distinction matters when you build ChatGPT apps.
If you are building a ChatGPT app, the interface question gets weird fast.
Do you design for a nice conversation, or for a system that actually does things on the user’s behalf? The answer is both, but not in the same way. Conversational UX is about the quality of the exchange. Agent experience is about what it feels like to delegate work, stay in control, and trust the system enough to let it act.
That distinction matters more than the name suggests. OpenAI’s own guidance for ChatGPT apps frames the useful part as know, do, and show: bring in new context, take real actions, and present information in a clearer UI than plain text (What makes a great ChatGPT app). That is already a hint that “good chat” is not the finish line.
The short version
If the user mostly wants to ask, clarify, and explore, you are designing conversational UX.
If the user wants to delegate, verify, and let the system take a real action, you are designing agent experience.
In practice, ChatGPT apps usually need both. The conversation is how the user expresses intent. The agent layer is how the app earns trust by turning that intent into something useful.
What conversational UX actually is
Conversational UX is the experience of using natural language as the interface. The user asks a question, gets a response, asks a follow-up, and the system stays coherent across turns.
That sounds simple, but the good version is more than chat bubbles. It has to handle ambiguity, corrections, missing context, and the fact that people rarely know exactly what to ask on the first try.
OpenAI’s app guidance points in this direction when it talks about helping users work with new context and present results in a clearer way than raw text. In other words, the conversation is not the product. It is the interaction pattern that reveals the product’s value (OpenAI, 2026).
That is why conversational UX is often strongest when the user is:
- exploring a problem
- narrowing down options
- asking follow-up questions
- trying to understand a complex situation
A good example is a support or research workflow. The user starts broad, then gets more specific. The interface should feel patient. It should not force them into a form before they know what they need.
What agent experience changes
Agent experience is the next layer. It is not just “chat, but with tools.” It is the experience of letting an AI system carry part of the task while the user stays oriented.
That means the app has to answer a few hidden questions:
- What is the system allowed to do?
- What will it change?
- How will the user know what happened?
- How do they recover if the system gets it wrong?
This is where the older HCI research gets useful. IBM’s work on conversational control of GUI found evidence of improved engagement and satisfaction, and even a significant increase in trust after users worked with the conversational UI (IBM Research, 2024). That is a strong reminder that agent-like interaction is not only about speed. It is about confidence.
I use AX here as shorthand for agent experience. It is not a formal standard term. It is just a useful way to describe the part of the product where the user is no longer merely talking to software, but delegating work to it.
That delegation is the whole game.
If conversational UX asks, “What should the user say next?”
Agent experience asks, “What should the system do next, and how do we make that feel safe?”
Why the difference matters in ChatGPT apps
ChatGPT apps sit right on the seam between these two ideas. OpenAI’s current app page makes that pretty explicit: apps can take action across tools, bring in live context, and keep users in control through permissions and workspace settings (Apps in ChatGPT).
So if you are building one, the question is not “Should this be conversational?”
The better question is “Which parts should be conversational, and which parts should feel like a trusted assistant?”
That is where a lot of teams get sloppy. They build something that can chat, but it does not really know, do, or show anything meaningful. Or they build something that does useful work, but the user never understands what happened.
OpenAI’s own checklist is a good filter here: does the app clearly add new things to know, new things to do, or better ways to show results? If the answer is no, the app is probably just a thin wrapper (OpenAI, 2026).
A simple way to separate the two
Here is the mental model I keep coming back to.
| Question | Conversational UX | Agent experience |
|---|---|---|
| What is the goal? | Help the user express intent naturally | Help the user delegate work safely |
| What matters most? | Clarity, flow, recovery, and follow-up | Trust, control, permissions, and outcomes |
| What should feel good? | The conversation | The handoff and the result |
| What can go wrong? | Confusion, dead ends, awkward prompts | Wrong actions, hidden state, loss of trust |
The table is the point. A lot of products say they are “conversational” when they really need to be trustworthy agents. Those are different design problems.
IBM’s 2010 CoCo paper is a useful historical marker here. It described a conversational interface that could synthesize a plan, execute it, and return the result to the user. That is basically the shape of modern agentic UX, just with better models and better product surfaces (IBM Research, 2010).
What each one is good for
Conversational UX is strongest when the user is still figuring things out.
That means:
- discovery
- narrowing
- explanation
- planning
- comparison
Agent experience is strongest when the user already knows the job and wants it done.
That means:
- research on their behalf
- pulling connected data
- updating records
- triggering workflows
- summarizing action back to the user
OpenAI’s current app examples make this split pretty clear. In the ChatGPT app directory, apps are framed as a way to complete tasks, bring in context, and get more done in the conversation. The examples lean heavily on action and access to live data, not just chat polish (OpenAI, 2026).
That is why the most useful ChatGPT apps usually start small. They do one job well. They do not try to mimic the entire product surface.
Where teams get it wrong
The first mistake is confusing a conversational UI with value.
You can make almost anything feel conversational. That does not mean it helps the user.
The second mistake is over-porting the product. OpenAI explicitly warns against cloning the whole product into ChatGPT. The better pattern is a focused surface with a few capabilities that map cleanly to real jobs-to-be-done (OpenAI, 2026).
The third mistake is hiding the agent work. If the app takes action but never shows what it changed, users lose the thread. That usually makes the product feel clever instead of useful.
The fourth mistake is confusing output with control. A summary is nice. A summary plus an action log, a permission model, and a recovery path is what makes the experience trustworthy.
How to design for both
This is the part I think teams should spend more time on.
Start with conversational UX for the entry point. Let the user talk naturally. Let them be vague at first. Let the system ask the next useful question.
Then design the agent layer around the handful of things the app should actually know, do, or show. That is the scope boundary.
OpenAI’s guidance is useful here too. It recommends picking a small set of capabilities, naming them clearly, and making sure the app can answer the first meaningful prompt without the user needing to learn a new workflow first (OpenAI, 2026).
That lines up with what we see in ChatGPT itself. Apps are supposed to help with work, creativity, research, and everyday tasks, while admins still control availability and permissions in business environments (Apps in ChatGPT).
So the practical design rule is:
- Make the first turn easy.
- Make the action explicit.
- Make the result visible.
- Make recovery simple.
If you do those four things, the interface stops feeling like a chatbot and starts feeling like an assistant the user can actually rely on.
A good litmus test
When I am reviewing a ChatGPT app idea, I ask one question:
Would the user notice if this app stopped working tomorrow?
If the answer is yes, you probably have a real agent experience.
If the answer is no, you may just have a clever conversation layer.
That sounds blunt, but it is useful. The best apps are not the ones that talk the most. They are the ones that create new context, do something real, and present the result in a way the user can immediately use.
That is also why this topic matters for drio. If you are building a ChatGPT app visually, you are not just designing screens. You are deciding where the conversation ends and where the agent begins. Those boundaries are the product.
Takeaways
Conversational UX is about how the user and system talk to each other.
Agent experience is about how safely and clearly the system can act for the user.
For ChatGPT apps, the best designs use conversational UX to capture intent and agent experience to deliver value. If you only optimize the chat, the app feels thin. If you only optimize the action, the app feels opaque.
The sweet spot is a focused capability that knows, does, or shows something the user could not easily do inside ChatGPT alone.
If you want the bigger picture on how these apps fit together, read What Are ChatGPT Apps?. If you want the build side, Building MCP Tools with Rich UIs is the next stop.


