Related: sources · notes · metadata · Published Pieces
Human Voice Is Evidence
Choir should never clone voices. A real recorded voice carries information that transcript and synthetic speech cannot replace.
A recorded human voice is evidence.
That is the principle.
Not “voice is engaging.” Not “voice is intimate.” Not “voice increases retention.” Those may be true, but they are secondary. The important point is that voice carries information beyond words.
A transcript says what was said.
A voice says how it was said.
The difference matters. Human beings are exquisitely sensitive to vocal leakage. We hear hesitation, confidence, uncertainty, contempt, warmth, fear, fatigue, searching, performance, improvisation, irony, calculation, and care. We hear when someone is reading. We hear when someone is bluffing. We hear when someone is trying to maintain control. We hear when a sentence costs them something.
Not perfectly. People misread voices. Culture and context matter. Manipulators can perform sincerity. But voice is still data.
This is why voice cloning is poisonous for serious media. It takes the authority of embodied speech and detaches it from the body that paid the cost of speaking. It lets synthetic text borrow the aura of presence. It turns voice into a style asset.
A recorded human voice is evidence. A cloned voice is costume.
Automatic radio should build its trust around this distinction.
The AI narrator can speak. It can summarize, transition, contextualize, compare, and guide. But it should not pretend to be human. It should not be overly emotional. It should not stutter theatrically. It should not sigh, laugh, flirt, or perform fake intimacy. The AI voice should organize.
Human voice should testify.
When a podcaster, politician, scientist, critic, founder, journalist, or ordinary user said something important, automatic radio should retrieve the actual clip when available. It should preserve the source, timestamp, speaker, context, and link to the full work. Then the narrator can explain why that clip matters, whether it supports or contradicts a claim, what prior work it connects to, and how later discourse used it.
This is more legitimate than the current clip economy.
Today, people cut clips from podcasts and speeches, post them to TikTok, Instagram, Twitter, or YouTube Shorts, and the platforms capture much of the value. Attribution is inconsistent. Context is often stripped. The original creator may get distribution, but usually without a real provenance or reward system.
Automatic radio can do it better.
It can clip less, cite more, link to the full work, preserve context, allocate value, and make the clip part of a structured argument rather than a viral fragment. The goal is not to replace the original work. The goal is to route listeners toward it when it matters.
This also creates a new opportunity for ordinary users.
If a user speaks a thought into Choir and chooses to publish it, that speech can become a voice-bearing artifact. It is transcribed, segmented, indexed, cited, and connected to the public graph. If months later a future radio stream needs that exact perspective, it can play the user’s actual words in the user’s actual voice.
Speak once. Be cited in your own voice.
That is not voice cloning. That is memory.
The no-cloning rule changes incentives. If people know their real voice may become part of the record, they have a reason to speak with care. They are not just prompting a model to generate polished text. They are entering public discourse as embodied speakers.
This makes automatic radio more than audio content. It becomes a medium of accountability and presence. It preserves the person without pretending the machine is one.