Skip to main content
 /  Around 10 minutes to read

The Interface Paradox

Why we need both chat and navigation.

We’re building for two different brains. The tech world is splitting into camps. Chat interface evangelists on one side, traditional navigation defenders on the other. But we’re trying to serve two fundamentally different cognitive modes with one interface.

The Interface Paradox

What We Mean by “Chat” and “Navigation”

By chat I mean conversational interfaces where users type or speak to an agent and get a response. Think of ChatGPT, Claude, customer service bots, or AI copilots in apps.

By navigation I mean structured pathways: menus, tabs, search bars, sidebars, cards, filters. The traditional scaffolding that lets users skim, jump, and browse.

These modes aren’t mutually exclusive. In fact, that’s the point. We need both. Our brains need both.

The Two-Brain Problem: A Deeper Look at Cognitive Modes

Back in the 1970s, Marcia Bates’s research on information-seeking behaviour introduced this beautiful concept called “berrypicking.” The idea was that we don’t gather information in one big sweep, but rather pick it up in bits and pieces, like berries from different bushes.

Her work showed that directed searching and exploratory browsing are fundamentally different behaviors, and they’re both necessary for how we process information.

Think about it. When you’re in directed search mode, you know what you want. You’re laser-focused, filtering out the noise, hunting for that specific answer. Your brain is working hard here, maintaining focus, holding your goal in working memory.

But exploratory browse mode works differently. You’re wandering, making connections, following interesting threads. You don’t know exactly what you’re looking for, but you’ll absolutely know it when you see it. It’s like walking through a bookstore versus searching for a specific ISBN on Amazon.

These modes require different cognitive resources, and you literally cannot be fully in both modes at once. Your brain has to switch between them.

Google Search figured this out early. The search box triggers your directed mode when you have a question, but then the results page switches you to browse mode with ten blue links, “People also ask,” and related searches. They’re managing your cognitive transition without you even noticing.

And Google’s new AI Mode goes even further. You can now toggle between traditional search and full conversational AI, giving you the same query with two different cognitive experiences. They’re not forcing you to choose one paradigm; they’re letting you switch based on what your brain needs in that moment.

The Chat Revolution: Promise vs. Cognitive Reality

Back in 2023, ChatGPT hit 100 million users in two months, according to a Reuters analysis. The appeal was obvious: just ask for what you want in plain English! No menus to learn, no conventions to master.

But there’s something nobody talks about…

The Hidden Cost of Conversation

Traditional interfaces rely on recognition. You see “File → Save As” and your brain goes “yep, that’s what I need.” It’s almost effortless because you’re pattern matching, not creating.

Chat interfaces require much more cognitive work. They need you to remember what you want to accomplish, translate that into words, guess what phrasing the AI will understand, keep track of the conversation, and connect responses back to your original goal.

That’s why you pause before talking to Siri, mentally rehearsing your query. It’s why voice assistants still feel awkward after all these years.

Context: Linear vs. Spatial

Sure, ChatGPT now has memory and Claude remembers your conversations and context. But conversational context and spatial context are wildly different beasts.

Conversational context is linear, like having a phone conversation about a blueprint. You can reference “what we talked about yesterday,” but you can’t see it while you’re talking about today’s changes. You have to scroll back, losing your place, breaking your flow.

Spatial context is parallel. Look at Photoshop: layers on the right, tools on the left, canvas in the middle, history tucked in the corner. Your peripheral vision is constantly processing information you’re not even consciously thinking about. Try doing that in a chat window.

Stripe gets this. They added AI chat to their docs, but they didn’t replace the docs. Developers bounce between both in the same session, using chat for questions like “How do I refund a payment in Node.js?” and documentation for understanding how Stripe’s entire payment flow actually works.

One is a specific task while the other is building a mental model. Different cognitive jobs need different tools.

The Architecture of Understanding

Traditional interfaces encode information in their structure, creating not just navigation but a cognitive map.

Look at this URL:

HTML
github.com/facebook/react/blob/main/packages/react-dom/src/client/ReactDOM.js

Without clicking anything, without any explanation, you know exactly where you are: GitHubFacebook’s stuffReact projectmain branchpackages folderreact-domclient sourcethat specific file.

This is ambient information, knowledge you absorb without trying. Chat interfaces have to explicitly tell you everything or wait for you to ask.

The Navigation Overwhelm Problem

Want to see what happens when traditional navigation goes too far? Log into Cloudflare.

Holy. Shit. 🤯

The sidebar has sections within sections within sections. DNS, SSL/TLS, Security, Speed, Caching, Rules, Network, Traffic, Build, and each of those explodes into more options. It’s like they took every possible feature and gave it equal visual weight.

New users are paralyzed. I know I was when I first started using their platform. And I still am most of the time. You literally cannot figure out where to start without documentation. The interface has become so comprehensive it’s incomprehensible.

This is traditional navigation eating itself.

The Peripheral Processing Advantage

Traditional interfaces do something chat can’t: they enable peripheral learning.

In VS Code, you might never click “Source Control,” but seeing it there tells you version control is integrated. That awareness shapes your mental model of what VS Code is, even if you never use that feature.

Chat interfaces can’t provide this peripheral awareness. Everything has to squeeze through the narrow pipe of conscious attention, eliminating peripheral learning, ambient awareness, and unconscious map building.

The Serendipity Engine: How Discovery Actually Works

Spotify gets the balance right. Discover Weekly uses fancy algorithms to create your personalized playlist, but they keep the browsable stuff too: playlists, genres, and the ability to spiral through artist → album → related artists.

Why? Because that browsing spiral creates what information scientists sometimes call “information encountering,” which is really just a fancy term for stumbling onto cool stuff you didn’t know you wanted.

Pinterest’s Guided Search nails this approach. You search “bedroom ideas” and get bedrooms, but also little filter bubbles: “modern,” “cozy,” “minimalist.” Click “modern” and new bubbles appear: “industrial,” “scandinavian,” “mid-century.”

You’re not just finding what you searched for; you’re learning the vocabulary to describe what you actually want. That’s huge 👀

The Vocabulary Problem

This is the dirty secret of chat interfaces: they assume you know what to ask for.

But how do you learn the word “pull request” without seeing it in GitHub’s interface? How do you know to ask about “artboards” if you’ve never seen them in Illustrator or Figma’s menus?

Traditional interfaces teach through exposure. Chat interfaces can’t teach what you don’t know exists.

The Hybrid Revolution: Products That Get It

Some products are figuring this out by building bridges instead of choosing sides.

  • Notion’s Q&A sits right beside their hierarchical structure. When you ask a question and get an answer, you also see breadcrumbs showing where that information lives in the hierarchy. Click those breadcrumbs and boom, you’re browsing. It’s not replacing navigation with chat but creating a conversation between them.
  • Linear’s command bar is even slicker. Hit Cmd+K and start typing to enter directed search mode, see suggestions appear as a transition state, browse through options in exploratory mode, then click one to return to your task. The whole flow happens in under a second, and your brain switches modes without even noticing.
  • Perplexity understood something crucial: answers create questions. Their citation cards aren’t just sources but rabbit holes waiting to happen. You get your answer to satisfy your directed brain, but those cards sit there, tempting you to explore and activating your browse brain.
  • Cursor keeps your code fully editable while chat helps alongside. The file tab stays visible and line numbers stay referenced, with the physical location anchoring the conversation. You can jump between editing and chatting without losing your place in either world.

The Near Future: Ambient Computing and Cognitive Coexistence

Apple Vision Pro is doing something wild by making information exist in three states:

  • Peripheral: Windows floating in your space, visible but not demanding attention
  • Focused: Whatever you’re looking at directly
  • Conversational: What you’re talking to

You can browse peripherally while asking questions verbally while focusing visually. It’s the first interface that doesn’t force you to pick a cognitive mode.

Why Humane Failed: The Cognitive Bottleneck

Humane’s AI Pin removed the screen entirely. The Verge review called basic interactions “frustrating,” and that’s being generous.

Without visual feedback, every single interaction required formulating a verbal query, waiting for audio response, parsing speech in real-time, and keeping everything in your head.

No visual persistence means you can’t offload anything to the environment. Your working memory becomes the bottleneck. It’s exhausting. No wonder it failed.

The Design Principles That Actually Matter

State Persistence with Visual Confirmation

Chat is ephemeral, but work isn’t.

Linear gets this right. When you create an issue through chat, it immediately appears in your workspace as a real object you can see and touch. The chat was just the creation method; the issue is what matters.

Too many tools generate content that only exists in the conversation, forcing you to copy-paste it somewhere else to make it real. That’s backwards. Every chat action should produce something tangible, something that exists outside the conversation.

Context Breadcrumbs Everywhere

Even in chat, people need to know where they are.

Cursor nails this by keeping the file tab visible during chat and displaying line numbers right there. You’re not having an abstract conversation; you’re talking about this specific code in this specific place.

This spatial anchoring is huge because it connects the abstract conversation to the concrete location. Without it, users float in cognitive space, uncertain where they are or what they’re actually discussing.

Query Formation Assistance

Don’t make people guess what to ask. Help them figure it out.

GitHub Copilot Chat does something smart by suggesting follow-up questions after each response. These aren’t random suggestions but the actual questions people typically ask next in this exact workflow. It’s like having someone who knows the territory walking alongside you.

Making Dead Ends Into Doorways

Every answer should be a beginning, not an end. Good implementations:

  • Hyperlink entities to full documentation
  • Turn explanations into interactive diagrams
  • Make lists expandable into galleries
  • Transform static answers into explorable structures

The point is to make chat responses into launching pads for deeper exploration. Never leave users at a dead end.

Mode Switching Should Be Obvious

Users should never wonder which mode they’re in. Use different background colors for chat versus browse, change typography for conversation versus documentation, and add transitions between modes so the switch feels intentional.

Mode errors are expensive, cognitively speaking. Make the current mode visible in peripheral vision, not just in the focused area.

The Ambient Information Layer

Not everything needs interaction. Some information should just exist.

Status indicators, progress bars, breadcrumbs, and timestamps should be visible without demanding attention. They’re cognitive anchor points that orient users without requiring conscious processing.

Apple’s Dynamic Island gets this perfectly. It’s just there, showing what’s happening, never demanding you look at it but always available when you do.

Predictive Intervention

Know when users need help before they ask for it. Watch for:

  • Repeated failed attempts
  • Circular navigation patterns
  • Long pauses in familiar interfaces
  • Sudden changes in speed

But timing is everything. Jump in too early and you’re annoying, too late and you’re useless. The intervention has to match the user’s cognitive state, not some arbitrary timer.

What This Means for Builders Right Now

Stop Picking Sides

Chat versus navigation isn’t a competition. Your users need both, often for the same task.

Build dual-access patterns where every major function works through both conversation and navigation. Stripe’s docs let you browse to the Payments API or ask “how do I process a payment in Node.js?” to reach the same destination through different cognitive paths.

Create bridges between modes. When chat gives an answer, make entities clickable to browse. When browsing gets overwhelming, offer “Ask about this” to switch to chat. Make the transitions feel natural, not forced.

Track the right metrics by measuring not just task completion but cognitive efficiency. How many mode switches did it take? Where did users get stuck? Did they pick the right mode for the task?

Design for Natural Transitions

People shift between directed and exploratory modes constantly, so stop fighting it.

Add breadcrumbs to chat responses, make chat expand into browsable docs, and let complex browse interfaces collapse into focused chat. Use animation to make transitions feel smooth, not jarring.

Watch where users quit, which is usually at forced mode switches where the interface doesn’t support their cognitive state. That’s your redesign priority.

Respect the Limits

Both chat and navigation have cognitive costs that need to be minimized.

For chat interfaces, don’t make users hold entire conversations in memory. Create persistent artifacts, suggest queries to reduce formulation burden, and keep visual anchors during conversation.

For navigation interfaces, use progressive disclosure, maintain information scent, enable keyboard shortcuts, and keep frequent items accessible without deep diving.

Build for Half Attention

Nobody’s giving your interface their full cognitive resources. They’re Slacking, listening to music, chatting with coworkers, or thinking about lunch.

Core functions should work with minimal attention, complex tasks need to be resumable, state should be visually obvious, and errors should be recoverable without remembering the full context.

Test your interface while distracted. Can you still use it? Can you pick up where you left off? That’s the real test.

The Metrics That Actually Matter

Forget traditional metrics. Track what reveals cognitive reality:

  • Mode switch frequency: How often do users flip between chat and browse?
  • Abandonment points: Where exactly do users give up?
  • Discovery rates: Are features found through exploration or only direct search?
  • Recovery patterns: How do users recover from dead ends?

Learning curves: How fast do users get efficient?

These metrics tell you if your interface works with human cognition or against it.

The Bottom Line

Companies that understand cognitive modes will win. Users will flock to products that feel “intuitive” without knowing why.

The why is simple: these products respect how human brains actually work. They don’t force humans to adapt to one interface paradigm. They adapt to how humans think.

We don’t need chat to replace navigation. We don’t need navigation to resist chat. We need interfaces that understand when we’re hunting and when we’re gathering, when we know what we want and when we’re figuring it out.

Build for both brains, or watch someone else do it better.

Because the future isn’t chat or navigation. It’s understanding that we’ve always needed both.