Comment by siim
2 days ago
Curious what made you think the backend uses LLMs for content generation?
To clarify:
1. transcription is local VOSK speech-to-text via WebSocket
2. live transcript post-processing has optional Gemini Flash-lite turned on which tries to fix obvious transcription mistakes, nothing else. The real fix here is more accurate transcriber.
3. backend: TypeGraphQL + MongoDB + Redis
The anti-AI stance isn't "zero AI anywhere", it's about requiring human input.
AI-generated audio is either too bad or too perfect. Real recorded voice has human imperfections.
No comments yet
Contribute on Hacker News ↗