{
  "title": "Articles/chatbots-aint-it",
  "caption": "Chatbots Ain’t It",
  "slug": "chatbots-aint-it",
  "tags": [
    "ai",
    "article",
    "artifacts",
    "chatbot",
    "choir",
    "draft",
    "hermes-published",
    "multiagency"
  ],
  "canonical_url": "https://mosiah.org/articles/chatbots-aint-it/",
  "interactive_url": "https://mosiah.org/#Articles%2Fchatbots-aint-it",
  "markdown_url": "https://mosiah.org/articles/chatbots-aint-it.md",
  "json_url": "https://mosiah.org/json/chatbots-aint-it.json",
  "fields": {
    "caption": "Chatbots Ain’t It",
    "created": "20260510182502150",
    "modified": "20260510182502150",
    "tags": "article hermes-published draft ai chatbot choir artifacts multiagency",
    "title": "Articles/chatbots-aint-it",
    "type": "text/vnd.tiddlywiki"
  },
  "text": "! Chatbots Ain’t It\n\n''Related:'' [[sources|Article Sources/chatbots-aint-it]] · [[notes|Article Notes/chatbots-aint-it]] · [[metadata|Article Metadata/chatbots-aint-it]] · [[Drafts|Drafts]]\n\nChatbots were the right way to discover that the model could talk. They are the wrong way to build the computer that comes after talking.\n\nThat sounds harsher than it is. Chat is still useful. I use it constantly. A good conversational agent can be a tutor, editor, rubber duck, analyst, search companion, therapist-shaped mirror, coding partner, and midnight sparring partner. The transcript is a powerful instrument for thought because language is the lowest-friction interface we have for intention.\n\nBut the industry has mistaken the discovery interface for the destination interface. The chatbot was the microscope slide on which we first saw the intelligence move. It is not the organism’s body.\n\nThe problem is not that chat is “text.” Text is great. The problem is that chat collapses too many distinct things into a single scrolling conversation: instruction, memory, state, evidence, output, revision history, delegation, tool logs, publication, identity, and collaboration. It gives one primary conversational agent one primary channel and asks it to impersonate an operating system.\n\nThat works for short exchanges. It breaks as soon as the work becomes social, durable, multi-step, or multi-agent.\n\n!! The transcript is not a workspace\n\nA transcript is a record of conversation. A workspace is a field of objects.\n\nThe difference matters because serious work is not a sequence of messages. Serious work accumulates structure. A reporting project accumulates sources, notes, claims, citations, versions, objections, interviews, drafts, edits, and published forms. A software project accumulates files, tests, issues, branches, logs, dependency graphs, design decisions, and rollback points. A design project accumulates sketches, components, assets, constraints, revisions, comments, and taste judgments. A video project accumulates clips, cuts, captions, scripts, b-roll, rights, renders, and timelines.\n\nNone of these are naturally “chat.” They can be discussed in chat. They can be initiated through chat. But their native form is artifact.\n\nThe chatbot interface hides this by making the model fluent enough to narrate work it has not structurally preserved. It can say “here’s the revised version” while the actual relation between versions is implicit in the transcript. It can say “I checked the sources” while the evidentiary graph is scattered across prior messages. It can say “we should make this a project” while the project does not exist except as a conversational intention.\n\nThe result is a strange UX bargain: the system feels intelligent moment to moment, then becomes amnesiac, uninspectable, and mushy when asked to sustain a world.\n\n!! Multiagency does not fit inside one mouth\n\nThe next mismatch is multiagency.\n\nAgentic systems are not one genius voice doing everything. They are increasingly ensembles: planners, retrievers, coders, reviewers, verifiers, browsers, sandboxes, indexers, schedulers, critics, and domain-specific workers. Even when the user experiences “one assistant,” the useful computation behind it is distributed.\n\nA chatbot UX has trouble representing that distribution honestly. If every subagent’s work is collapsed into one assistant message, the user loses provenance. If every subagent speaks directly in the transcript, the user gets spammed by process. If tool calls are hidden, trust erodes. If tool calls are exposed raw, the interface becomes a log viewer. If the main agent summarizes everything, the summary becomes another lossy bottleneck.\n\nThis is not merely a UI annoyance. It is an ontological error. Multiagent work needs shared state, role boundaries, handoff artifacts, partial outputs, confidence signals, reversible decisions, and inspectable traces. A conversation can coordinate those things, but it cannot be the only place they live.\n\nThe user should not have to “manage agents” any more than a video editor should have to manage individual CPU instructions. The agents should be internal organs of a larger machine. What the user should see are the objects those agents are transforming: the draft, the dataset, the claim graph, the cut, the pull request, the app, the map, the evidence bundle.\n\nAgents are not the interface. Artifacts are the interface.\n\n!! Multiplayer AI makes the failure obvious\n\nThe chatbot mismatch gets worse when more than one human enters the system.\n\nMost useful work is already multiplayer. Writing is shaped by audiences, editors, sources, collaborators, adversaries, and communities. Software is shaped by maintainers, reviewers, users, issues, CI, dependencies, and production incidents. Research is shaped by citations, replication, institutional memory, disagreement, and priority disputes. Politics is shaped by coalitions, media surfaces, oppositional frames, and public records.\n\nA private chatbot thread is the opposite of that. It is a sealed dyad: one user, one assistant, one transcript. Even when the model has tools, the social shape remains “I talk to my bot.” That is a poor substrate for multiplayer intelligence.\n\nReal social AI needs objects that multiple people and multiple agents can point at together. It needs claims with provenance. It needs comments attached to passages, not buried three scrolls later. It needs versions that can be compared. It needs citations that can be audited. It needs permissions, ownership, attribution, publication states, and rollback. It needs memory that is not just “what the assistant remembers,” but what the group has made true enough to preserve.\n\nChat is excellent for saying “look at this.” It is bad at being the “this.”\n\nThe empirical proof is already in front of us: people almost never share and read other users’ chat threads as content. A brilliant conversation may be personally useful to the person who had it, but the transcript rarely becomes a public object other people want to consume. Shared ChatGPT threads feel like overhearing somebody else’s session notes. They are context-heavy, rhythmically awkward, and almost always too long for the value they contain.\n\nGroup chat is worse. Even in group chats with humans we love, we rarely scroll back up and read the log as literature, argument, documentation, or institutional memory. The log is socially alive in the moment and nearly dead afterward. That is not a viable content format. Adding bots to the room does not fix the form; it intensifies the entropy.\n\nThis is why social AI cannot simply be “group chats with bots.” Group chat already has enough entropy. Adding agents to the room makes it worse unless the agents are oriented around shared artifacts. Otherwise every bot becomes another participant generating text into the same channel, competing for attention, performing competence, and losing the thread.\n\nThe solution is not more chat. The solution is a shared artifact layer where conversational activity becomes marginalia, modulation, command surface, or temporary scaffolding around durable objects.\n\n!! The proactive chatbot is the reductio\n\nThe failure becomes almost comic in proactive AI.\n\nIf a chatbot wakes up and sends you a message, what should it say? In theory, something useful. In practice, the obvious failure mode is contextless relevance theater: pseudo-work, pseudo-code, pseudo-insight, or recommendations detached from the living projects where they would matter.\n\nThis is why a morning “briefing” full of hallucinated pseudocode feels so alienating. The issue is not simply that the code is wrong. It is that code without project-local context is barely an object. It has no repo, no tests, no design pressure, no user story, no surrounding architecture, no failing bug, no committed intention. It is not work. It is the smell of work.\n\nA useful proactive system would not wake you up to generic fragments. It would operate against artifacts you actually care about. It would say: the draft has a weak transition here; this source contradicts paragraph four; the build failed after this dependency changed; your claim about managed agents needs a citation; these three notes want to become one piece; this video segment belongs in the published article; this old argument has become newly relevant because of today’s news.\n\nThat kind of proactivity requires durable state. It requires taste, memory, source graphs, project boundaries, and verification loops. It requires the system to know what counts as progress inside a particular artifact.\n\nWithout artifacts, “proactive AI” becomes a chatbot knocking on your door to perform intelligence.\n\n!! Why files helped coding agents\n\nCoding agents worked earlier than many other agents because code already has a decent artifact substrate.\n\nA repo is not just a pile of text. It is files, tests, commits, diffs, branches, issues, reviews, CI, package constraints, runtime errors, and deployment state. Git makes mutation reversible. Tests make some claims executable. Diffs make changes inspectable. CI makes verification social and mechanical.\n\nThis is why the “Claude likes files” lesson is too small. Files are not magic. The magic is branchable, inspectable, reversible, provenance-bearing state.\n\nFor code, we inherited that substrate. For everything else, we mostly have chat transcripts, cloud docs, feeds, folders, and vibes.\n\nThe next computer needs Git-like properties for more than code. Not literally Git everywhere, but the same civilizational affordances: version, diff, branch, merge, cite, verify, rollback, publish, fork, annotate, sign, and remember.\n\n!! The automatic computer\n\nThe endpoint is not a better chatbot. It is an automatic computer.\n\nAn automatic computer is not a device an AI “controls” on your behalf. That framing is already too small and too creepy. People do not want a ghost moving their mouse across a personal desktop. They have multiple devices, intermittent attention, privacy boundaries, and an ownership relation to their tools. Watching an agent puppeteer a GUI is useful as a bridge, not as the final form.\n\nThe automatic computer is a persistent backend runtime attached to durable artifacts. Text and voice can still be the easiest inputs. Chat may still be the social command line. But the outputs are not assistant messages. The outputs are living objects: articles, source bundles, claim graphs, apps, videos, notebooks, datasets, dashboards, CAD models, playlists, games, workflows, agents, and publications.\n\nThe runtime handles multiagency internally. It can spawn workers, run checks, preserve traces, route models, schedule jobs, fetch sources, maintain memory, and propose edits. But the user’s world is organized around artifacts, not agent personas.\n\nThe open UX question is how much conversational scaffolding should remain visible. One possibility is artifacts plus chat: the user talks beside the object, gives instructions, leaves comments, modulates tone, and asks for transformations. Another possibility is closer to chatless vtext: the user edits the artifact directly, and input can be either content or meta-instruction depending on context. “Make this sharper” and a rewritten sentence both happen at the site of the text. Each revision becomes a new version. The user can cycle through versions, compare them, roll back, and feel the object changing under their hands rather than watching an assistant narrate changes from across the room.\n\nThat may be the deeper break. The right interface might not be “chat around artifacts” so much as “artifacts that can interpret local intent.” The user modifies the text; the system infers whether the input is content, instruction, constraint, or taste signal, and processes it smoothly. It is not seamless yet. But the feeling matters: direct manipulation of a living artifact, with versions as the conversation’s memory.\n\nThis also changes what “memory” means. In a chatbot, memory is often private personalization: the assistant remembers your preferences. In an artifact-native system, memory is public enough to be useful: this claim came from that source; this revision replaced that paragraph; this app was generated from that spec; this citation was disputed; this note later became this published piece. Memory becomes provenance, not vibes.\n\n!! Chat remains, or maybe it dissolves\n\nNone of this means chat simply disappears. But it does mean chat loses its claim to be the primary content form.\n\nChat is powerful because it is the universal ingress for ambiguity. It is where intention forms before it has structure. It is where a user can say, “this is wrong,” “make it sharper,” “what am I missing,” “turn these notes into an article,” or “ship the simple version.” It is also where social texture lives: jokes, frustration, taste, doubt, disagreement, persuasion.\n\nBut maybe chat should be the porch, not the house. Or maybe, for some classes of work, even the porch disappears into the artifact.\n\nThe uncertain boundary is important. One future is artifact-centered chat: comments, instructions, voice notes, and conversational turns modulate a text, app, dataset, video, or claim graph. Another future is chatless or near-chatless: the user works directly in the artifact, and the system treats edits, selections, marginal notes, and local commands as a blended input stream. The distinction between “content” and “instruction” becomes contextual. The artifact becomes the interface.\n\nThis is already visible in coding agents. If I set a coding agent to work for eight hours, I do not want to return and review a chat log — especially not a reverse-chronological or interleaved tool log. I want the agent to produce an artifact: a Markdown, PDF, or HTML report explaining what changed, why it changed, what remains risky, which tests ran, where the diffs are, and what decisions require human review. The work happened over time, but the content format for review is not the log of that time. It is the synthesized artifact.\n\nThe house is the artifact system: the place where work survives the conversation, where multiple agents can operate without spamming the user, where multiple humans can collaborate without losing the object, where evidence can be audited, where revisions can be compared, and where finished things can be published.\n\nThis is the actual break from the chatbot era. Not “AI that talks better.” AI that leaves better objects behind — and perhaps lets us work on those objects directly enough that chat becomes only one input mode among many.\n\n!! The definitive line\n\nChatbots made AI legible by making it conversational. That was necessary. It was also distorting.\n\nThe future is not many smarter characters in a chat window. The future is multiagent computation organized around shared, durable, inspectable artifacts. The social form is not “everyone talks to a bot.” The social form is humans and agents working on the same objects, with provenance, revision, verification, and publication built in.\n\nI am less certain about the final input form. It may be chat modulating artifacts. It may be comments and commands attached to artifacts. It may be vtext: direct artifact editing where user input can be content, instruction, taste signal, or revision request, and every transformation is a navigable version. But the negative claim is now obvious.\n\nChatbots ain’t it because conversation is not the durable unit of work, review, collaboration, or publication.\n\nArtifacts are.\n"
}