Instant code, endless review
Claude’s voice mode is making code faster than humans’ ability to review it.
Type /voice, hold the spacebar, and speak. That’s all it takes to speak software into existence using Claude Code, Anthropic’s AI coding agent. It’s capable of generating, editing, and debugging software in real time.
Unlike GitHub Copilot and Cursor, Anthropic’s voice transcription tokens are free. Early users report that the experience is genuinely fluid: spoken instructions translate cleanly into code changes, and iterating conversationally feels faster than typing prompts. Anthropic was reached for comment regarding the long-term pricing of voice tokens, but the company did not respond by the time of publication.
Voice mode is being promoted as a way to further reduce friction between human intent and AI execution. It’s the pinnacle of “
vibe coding,” a term coined by AI researcher Andrej Karpathy to describe a shift from writing code line-by-line to high-level description. Enthusiasts praise it for speed and fluidity. But beneath the excitement, a more complicated picture is taking shape.
Security gaps are a common issue emerging in projects created by “vibe coding” that appear polished.
A major debate around this first hit developer circles after veteran developer Joel Reymont used Claude Code to generate 13,000 lines of code for the OCaml compiler in just a few days. He saw it as a massive leap in productivity. However, the OCaml maintainers rejected the contribution, arguing that the sheer volume of AI-generated code created an impossible and exhausting review burden.
It’s an extreme example of a more pervasive problem that developers are encountering, where the speed and amount of AI code is not yet at a point where it can stand without the manual work of humans. Engineers argue that the process of checking AI code can be even more costly and time-consuming than simply writing the code by hand. One developer I spoke to who is familiar with AI coding tools said the technology is improving rapidly but still requires experienced engineers to guide it.
While the ability for AI generated code to run has jumped from 20% to over 90% in just two years, it isn’t necessarily safe. A Veracode analysis revealed that 45% of AI-generated snippets contain security vulnerabilities, a rate that matches human error but at a volume humans can’t possibly audit. The risk is less about the frequency of bugs and more about the time cost of debugging complex, AI-written code that often lacks proper architecture.
Those gaps have had real-world consequences. In early 2025, the founder of a data-enrichment startup called Enrichlead boasted on X (formerly Twitter) that the platform had been built with “zero hand-written code” using the AI coding tool Cursor. Within days, users discovered major security flaws, including the ability to bypass paid features and manipulate data. The vulnerabilities quickly drew criticism across developer and security communities, and the project was ultimately abandoned after the founder acknowledged he lacked the expertise to audit and secure the roughly 15,000 lines of AI-generated code.
Reviewing that much code is a massive undertaking, yet managers and project leads increasingly expect it from senior developers.
One detailed account on r/ClaudeCode described the day-to-day experience as “babysitting a system that actively resists completing work,” with the cognitive burden of constantly verifying AI output is more exhausting than writing the code manually.
Others worry about what this means for the engineers coming up behind them. “Asking it to do something you don’t know how to do, that’s just asking for a disaster,” one developer warned on r/ClaudeAI.
The shift is already underway. A LeadDev survey found that 54% of engineering leaders plan to hire fewer junior developers. AI will now be handling the entry-level tasks that once trained senior talent, leaving a gap in the workforce of experts capable of catching subtle hallucinations and logic errors.
Critics fear that this reliance on AI coding could create a dangerous precedent where founders choose immediate speed over long-term stability. Founders may be left with platforms that are un-auditable and un-scalable.
The engineers most bullish on AI coding tools tend to be the ones who are also clearest about their limits. Simon Willison, a prominent developer and blogger who has written extensively about LLM capabilities, has described the current moment as an inflection point.
What the responsible hybrid looks like, based on the emerging consensus in developer communities: AI for acceleration on well-understood, bounded problems; human review before anything enters a shared codebase, and ongoing investment in junior developer mentorship.
The OCaml episode, in the end, may be the most instructive data point. Joel Reymont wasn’t wrong that the problem was worth solving. What the maintainers pushed back on was the assumption that working was enough and that the community would be able to absorb those 13,000 lines of AI-generated code and stand behind it indefinitely.
Voice mode may make it faster to get to work, but it doesn’t change what comes after.



