{
  "title": "Articles/audio-runway-buys-cognition-time",
  "caption": "Audio Runway Buys Cognition Time",
  "slug": "audio-runway-buys-cognition-time",
  "tags": [
    "agents",
    "article",
    "automatic-radio",
    "hermes-published",
    "pack-10",
    "published",
    "work-radio"
  ],
  "canonical_url": "https://mosiah.org/articles/audio-runway-buys-cognition-time/",
  "interactive_url": "https://mosiah.org/#Articles%2Faudio-runway-buys-cognition-time",
  "markdown_url": "https://mosiah.org/articles/audio-runway-buys-cognition-time.md",
  "json_url": "https://mosiah.org/json/audio-runway-buys-cognition-time.json",
  "fields": {
    "sort-date": "2026-05-12T13:25:00Z",
    "caption": "Audio Runway Buys Cognition Time",
    "created": "20260512130026429",
    "modified": "20260512130026429",
    "tags": "article hermes-published published work-radio automatic-radio agents pack-10",
    "title": "Articles/audio-runway-buys-cognition-time",
    "type": "text/vnd.tiddlywiki"
  },
  "text": "//Related:// [[sources|Article Sources/audio-runway-buys-cognition-time]] · [[notes|Article Notes/audio-runway-buys-cognition-time]] · [[metadata|Article Metadata/audio-runway-buys-cognition-time]] · [[Published Pieces]]\n\n! Audio Runway Buys Cognition Time\n\n//The graph gives the radio runway. The runway gives agents time. The agents update the graph.//\n\nThe central advantage of automatic radio is temporal.\n\nIn chat, latency is dead time. The user asks a hard question, then waits. If the model thinks for three minutes, the product feels broken. If a research agent needs thirty minutes, the user leaves. If a coding agent needs four hours, the user cannot sensibly remain in the thread.\n\nAudio changes the economics of waiting.\n\nIf the system has an artifact graph, citation network, prior vtexts, source clips, summaries, human recordings, claim maps, and cached transforms, then a single prompt can produce immediate runway. The system can begin with what is already known while deeper agents work in the background.\n\nThat runway buys cognition time.\n\nA user asks to catch up on the voice AI landscape and what it means for Choir Radio. The system does not need to wait until every research path finishes. It can start with orientation: realtime audio models, cascaded STT-text-TTS pipelines, and whether automatic radio is better understood as a voice agent or as media traversal over an artifact graph.\n\nThen it can traverse existing material: the current position that voice should be thin, intelligence should remain in text-native multiagent systems and artifact graphs, and realtime presence should not own memory.\n\nNow background agents have time. One searches open-source realtime voice models. Another compares STT and TTS components. Another reviews competitors. Another drafts an implementation plan. Another extracts the cost model. The radio keeps going.\n\nA few minutes later, the research agent updates the stream: public evidence still favors hybrid pipelines for reasoning-heavy tasks. Native speech systems are improving, but they often trade depth for immediacy. This supports the current Choir architecture.\n\nThe user never stared at a spinner.\n\nThis is the product insight. Automatic radio can make long inference feel alive because the user is already receiving value while the system computes. The listening stream becomes a buffer between human attention and machine time.\n\nThis only works if there is reusable cognition. A shallow voice agent cannot do it because every answer has to be generated from the current conversation. A strong artifact system can do it because the graph already contains paths: prior work, related claims, source trails, objections, human clips, cached summaries, known distinctions.\n\nThe richer the graph, the longer the runway.\n\nA mature Choir system should be able to speak for thirty minutes on a serious topic using existing structure while new agents deepen the answer. That does not mean rambling. It means moving through layers: orientation, context, sources, prior art, disagreement, synthesis, unresolved questions, and current work.\n\nThe user can interrupt at any point: skip the background, go deeper on the benchmark, play the original clip, what did I say about this last week, how does this affect the build. The interruption becomes an event. The radio branches, answers, and returns.\n\nThis makes audio different from text. Long text output often burdens the user. Long audio, if well produced and interruptible, can be the ideal medium. People already listen for hours to podcasts, interviews, lectures, and radio. The problem is not duration. The problem is signal.\n\nAudio runway works when the stream remains source-grounded, paced, and responsive. It fails when it becomes filler. The system must always know what it is doing: orienting, explaining, contrasting, quoting, updating, or asking for a decision.\n\nThe background agents should not be hidden in the sense of unaccountable. Their results should land in artifacts. The radio may summarize, but the artifact records. The user can later inspect the sources, vtext, code diff, checkpoint, or decision record.\n\nThat is the full loop: the graph gives the radio runway; the runway gives agents time; the agents update the graph; the radio reports what changed; the user steers; the artifact remembers.\n\nThis is why automatic radio is not merely an audio interface. It is the temporal layer of an automatic computer.\n"
}