← Stream
articles

Is conversation still a useful metaphor for human-machine interaction?

A collection of thoughts & ideas on alternative metaphors for human-GenAI-interaction.

Is conversation still a useful metaphor for human-machine interaction? Loose thoughts

Tended March 21: added sections on the problem with conversation and delegation as metaphor.

Before generative AI

  • chatbots and voice assistants were the primary conversational interfaces
  • primary tasks: answering questions and self service, primary use cases: mostly CX and marketing
  • rigid interaction model: turn taking/input-processing-output
  • interaction metaphor closest to human experience: conversation/turn-taking
  • goal: reduce friction in an experience that’s so cloinky that no-one can be tricked into thinking this is a human.
  • rationale: the more the interaction between human and machine resembles human-human conversation, the easier it gets to talk to a machine, the more people will trust the interface.

After generative AI

  • chatbots become a generic name for all kinds of tasks and use cases: answering questions, performing tasks, generating content, workflow management, coding
  • different use cases require different input types: traditional question & answer/turn taking, but also simple data ingestion, or command line inputs.
  • and different types of output: often still answers to questions, but also deliverables, end results of a workflow: a summary, a document, a video, an image. That’s not part of a conversation per se. Unless you consider a conversation about, say, holidays where someone shows you holiday pictures. Or if you consider your manager asking for a report and you handing over that report a few days later, a conversation. That’s really stretching things.

Generative AI as a mimicry of human conversation

  • Large Language Models don't contain language
  • General purpose chatbots like ChatGPT don’t have agency, consciousness, an inner locus. Yet the surface form takes the shape of something that is familiar to humans: conversation. Without fullfilling the underlying contract that underpins human-human conversation:
  • grounding in truth, shared sense making, sensory experience, deixis
  • so even in chat interfaces, the interaction happening is not a proper conversation.

The problem with conversation

Conversations between humans rely on a lot more than two-way interaction: having equal stakes in the exchange, relying on an underlying contract with a lot of subtleties and intricacies, and a lot of non-verbal communication, implicature, and assumed trust.

Calling a human interaction with a chatbot a conversation has always been a bit of a misnomer, but with so many different use cases opening up with GenAI, we’re really stretching that metaphor beyond usefulness.

There’s also the anthropomorphism problem: conversation is inherently human-shaped. For many GenAI use cases, you don’t want that. You don’t need your code editor to feel like a person. You don’t need your document processor to have a personality. Yet the conversation metaphor smuggles in anthropomorphism by default.

And then there’s the emotional labor dimension: when we design chat agents that converse, we’re also designing agents that perform emotional work. See Helena Rodemann’s Personality is a labor system on emotional labor by chat agents.

Metaphor: delegation and delivery

Imagine having a large document set, and you need to perform a number of actions on it. With agent AI, yes, you might be inputting things through language, but chances are that most of the time, you’ll be using a mix of language, traditional GUI elements like forms, upload fields etc.

After that, chances are that a small swarm of agents will take your request and run with it, leaving you standing with no clue of what’s happening, and then returning you your requested deliverable. That’s not a conversation anymore, that’s more of a parcel delivery service :-) Of course, you could argue that after that, you could use turn taking to tweak the results, but nine out of ten, you’d be doing that directly in the deliverable, not through conversation.

Alternatives to talking

Instructing, Commanding, Inputting, Entering, Prompting, Directing, Delegating, Dictating, Querying, Requesting, Dispatching, Steering, Cueing

Who do we talk to?

When we talk to a chatbot, does it help us to think we’re talking to something else?

Instructing: asking someone else to do something. Out of your hands, out of your brain.

What is the voice that we hear in our head when we chat with a textbot? Do we hear someone else’s voice, or our own? And does it matter?

And who’s listening?

NL: luisterbereidheid: the willingness to listen. Listening is intentional, silence is not. People can be not luisterbereid. Organisations too. LLMs too? If there’s no luisterbereidheid, ideas, plans and changes cannot land.

How do you know someone’s listening? It’s also different than hearing. Listening is receptive, like an embedding.

Systems can also be listeners. Listeners poll and ping. Receiver and listener are different.

Threadiness

On LinkedIn today (April 1, 2026): Threadiness https://www.linkedin.com/posts/gcadorette_glengorithms-share-7445071189846167552-5d4b?utm_source=share&utm_medium=member_desktop&rcm=ACoAAADZio0B3fvibsQhqorZYUThaK9aae5bgWs

Every reply in a real conversation does 3 things:

1️⃣ Semantic expansion: Each response adds new tokens to your post’s topic map. The NLP engine uses them to sharpen who else should see it. The thread is the algorithm learning who your people are.

2️⃣ Session persistence: People stayed, replied, and came back. That’s rare behavior. And rare behavior gets amplified.

3️⃣ Low friction amplification: It’s cheaper for the algorithm to keep pushing a post that’s already proven it can hold attention than to keep testing one that people nodded at and kept scrolling.

This is threadiness.

It’s a trailing indicator of content that had somewhere to go.

Depth can be gamed too. Anyone who’s watched a “first!” comment turn into a 6-reply nothing-thread knows exactly what data-poor engagement looks like.

Maybe that’s why LinkedIn decided to prioritize depth over volume.

My theory? It’s a response to manufactured engagement. Pods, reciprocal commenting, performance metrics. All that good stuff.

By prioritizing thread depth over comment volume, the algorithm can start distinguishing between a real community having a discussion and a group of people just showing up for each other’s numbers.

Bacteria to AI: WOlfe (2024) jagged ontologies:

creating a densely textured conceptual landscape of high dimensionality that never resolves to a smooth homogeneous surface.

Metaphor: thinking & inner dialog

What if we build an interaction model metaphor around thinking? With an audience of one: ourselves? What would such an interface look like? Assuming it’s language based, a thought-centered interface would be very minimal, with no distractions, so you can focus on capturing thoughts and giving words to them. Thinking activates us, invites us to explore and expand mental models and to find words for new ideas. It helps us reflect on emotions, take actions, keep track, and discern true from false. All this is something that we typically do alone, in an inner dialogue. Not in a conversation.

Metaphor: note-taking and meta-cognition

Note taking is an interesting concept in this light: it’s what we do as a connected aside next to our main thinking or task: taking notes while talking, while reading, while studying. In this garden, I write on an empty canvas. Nothing there, not even a menu bar. But while writing, I also use AI for note-taking and task execution. Not as a conversation with another thing, but as a stream-of-thought note while writing: research this, create a new card for that, update my project in the background while I keep writing here. I’m not moving to the page margin or a separate entity for note taking, but do it right here, in my task domain.

Front-end back-end doesn’t really cover this, it’s front-end front-end, or perhaps core-marginal. Also the notes become executionable.

Research papers

Conversational grounding and its absence in AI:

  • Clark & Brennan, Grounding in Communication (1991): the foundational framework for how conversational participants establish mutual understanding
  • Shaikh et al., Grounding Gaps in Language Model Generations (2023): LLMs produce fewer grounding acts than humans and presume common ground
  • Suchman, Plans and Situated Actions (1987/2007): machines designed on planning models miss the situated nature of human communication

The conversation metaphor and its limits:

Alternative metaphors:

Thinking and metacognition as interaction paradigms:

  • Thought as a Substrate in Human-AI Interaction (2024, ACM UIST): proposes treating thought itself as the primary substrate for human-AI interaction, arguing that interfaces should be built around cognitive processes rather than conversational turn-taking
  • Interacting with Thoughtful AI (2025): proposes a paradigm where AI functions as a continuously thinking entity rather than a passive responder, enabling proactive collaboration through surfacing intermediate thoughts, full-duplex processing, and shared cognitive workspaces
  • Tools for Thought (2025, CHI workshop): synthesises research from 34 papers on how GenAI expands automation while simultaneously creating both risks and opportunities for human cognition — memory, creativity, and critical thinking
Mycelium tags, relations & arguments