▶️ LISTEN TO THIS ARTICLE
Collins Dictionary named "vibe coding" its word of the year for 2025. Andrej Karpathy coined the term in February of that year to describe building software by talking to an AI, accepting whatever code it produces, and trusting the vibes. Within months, the joke became mainstream practice. GitHub Copilot crossed 20 million users. Cursor grabbed significant market share. The vibe coding tools market hit $4.7 billion. And now, barely into 2026, the disillusionment reports are stacking up faster than the pull requests.
Retool, O'Reilly, Stack Overflow, Palo Alto Networks, Red Hat, and Veracode have all published critical assessments in the past three months. The pattern is consistent: vibe coding delivers genuine speed on small projects and prototypes, then falls apart in ways that are expensive, insecure, and difficult to reverse.
The Security Numbers Are Bad
Veracode's 2025 GenAI Code Security Report tested code from over 100 large language models across Java, JavaScript, Python, and C#. The finding: AI-generated code introduced security vulnerabilities in 45% of test cases. When given a choice between a secure and insecure method, the models chose the insecure option nearly half the time. Java was worst at 72% failure. And the kicker: security performance stayed flat regardless of model size or training sophistication. Bigger models wrote more functional code. They didn't write safer code.
Security startup Tenzai ran a head-to-head assessment of five major vibe coding tools in December 2025, including Claude Code, OpenAI Codex, Cursor, Replit, and Devin. They built three identical test applications with each tool. Result: 69 vulnerabilities across 15 applications, roughly half a dozen rated critical. The tools handled generic security patterns well enough. They failed where safe code depends on context, which is where real security actually lives.
Then there's Lovable, the vibe coding platform that reached unicorn status by letting anyone build full-stack apps through chat. Security researchers scanned 1,645 apps from its showcase directory. 10.3% had critical flaws exposing user data through misconfigured database policies. Names, emails, API keys, payment details, personal debt amounts. The apps worked. They just leaked everything.
Palo Alto Networks' Unit 42 team published their own advisory on vibe coding security, proposing the SHIELD governance framework for organizations adopting AI coding tools. Their assessment warns that current vibe coding workflows lack the guardrails needed to catch security flaws before deployment.
Veracode's broader testing reinforces why: LLMs failed to defend against cross-site scripting in 86% of cases and log injection in 88%. Over 40% of AI-generated code contains security flaws, with missing input sanitization as the most common. The tools produce code that compiles, passes tests, and ships with vulnerabilities baked in.
The Maintenance Wall
Security is the urgent problem. Maintenance is the slow-moving one.
OX Security analyzed over 300 repositories in 2025, including 50 using AI coding tools. Their report, titled "Army of Juniors," identified ten anti-patterns present in 80-100% of AI-generated code: incomplete error handling, weak concurrency management, inconsistent architecture, excessive comments, and monolithic structures. The core finding wasn't that AI-generated code is more vulnerable per line. It's that vulnerable systems now reach production at unprecedented speed because review can't keep pace with output.
Builder.io documented an 8-fold increase in code duplication within AI-generated projects compared to traditional development. AI generates different patterns for similar problems, even within the same conversation. Ask for a data-fetching function on Monday and you get async/await. Ask for something similar on Wednesday and you get promise chains. Context windows mean the AI forgets architectural decisions from earlier sessions, so consistency degrades as projects grow.
Red Hat's analysis, published today, describes a "three-month wall" where vibe-coded projects hit sustainability collapse. The codebase grows beyond anyone's ability to maintain it mentally. Debugging becomes what one developer called "whack-a-mole": the AI fixes one thing and breaks ten others. Without specifications, the code itself becomes the only source of truth for what the software does, and code is terrible at explaining why it does what it does.
Forrester projects that 75% of technology decision-makers will face moderate to severe technical debt by 2026, up from 50% in 2025. First-year costs with AI coding tools run 12% higher than traditional development when you account for 9% code review overhead, a 1.7x testing burden from increased defects, and 2x code churn requiring constant rewrites. By year two, unmanaged AI-generated code drives maintenance costs to four times traditional levels.
Where It Actually Works
Here's the part the backlash pieces sometimes skip: vibe coding is genuinely productive for certain use cases, and dismissing it entirely misreads the situation.
Personal projects, prototypes, and MVPs benefit from the speed. A Red Hat developer described building five concept prototypes and three MVPs in months using vibe coding. Flashcard apps with flip animations and persistent storage, built entirely through prompts. For software that doesn't need to survive contact with production, scale, or compliance requirements, the productivity gain is real.
Linus Torvalds put it plainly at the Linux Foundation Open Source Summit. He called vibe coding "fine for getting started" and a "horrible, horrible idea from a maintenance standpoint." He'd used Google's Antigravity AI to vibe-code a Python visualizer for a hobby project, then hand-wrote all the C components himself. The distinction matters. Torvalds isn't anti-AI. He's drawing the line exactly where the data says it should be drawn: between disposable code and code someone else has to maintain.
Senior developers with deep codebase knowledge extract genuine value from AI tools. They spot bad suggestions, reject wrong patterns, and use the speed boost on tedious tasks while maintaining architectural coherence. The METR study found experienced developers actually got slower with AI on their own repos -- precisely because they were already so proficient on familiar codebases that AI assistance added overhead rather than value. For routine boilerplate, the tools save real time.
The problem isn't that vibe coding fails everywhere. It's that the failure modes are invisible until they compound, and the use cases where it works don't look like the use cases where organizations are deploying it.
What Comes After the Vibes
The industry isn't abandoning AI-assisted development. It's trying to figure out what structured AI-assisted development looks like.
O'Reilly's Signals for 2026 report frames the shift: enterprises are moving from experimentation to accountability, with the fundamentals of building and maintaining good software becoming what separates purposeful AI-assisted code from the crowd. Philip Guo's concept of "vibe checks" proposes simpler verification scripts that validate AI output without reviewing every line, like checking file size ratios to catch silent data loss.
Amazon launched Kiro, GitHub released Spec Kit, and startups like Codeplain and Tessl are building tools around spec-driven AI development. The premise: write specifications first, then let the AI implement against testable constraints rather than open-ended prompts. The approach treats AI as a fast typist working from blueprints rather than an architect working from vibes.
Stack Overflow's January 2026 experiment with vibe coding told the story in miniature. A non-developer used Bolt to build a bathroom review app in ten minutes. It looked finished. Then developer friends reviewed it: no security features, inlined styling, no unit tests, messy repository structure, buried components. Fixing the fundamentals required exactly the junior developers that vibe coding was supposedly going to replace.
The emerging pattern is a two-phase workflow: vibe to explore, then specify to build. Use natural language to prototype rapidly. Once the concept works, write actual specifications and constraints before scaling. As one developer put it, if you can write a unit test to validate the output, the scope is small enough to vibe. If you can't test it at that level, you need a spec.
That's not a failure of AI-assisted development. It's the practice growing up. The vibe coding hype promised that talking to a chatbot could replace understanding software. The backlash is proving that wrong. What's actually emerging is something more useful: AI as a fast implementation layer that still requires human judgment about architecture, security, and long-term maintainability. The AI coding productivity paradox has always pointed here. Individual speed was never the bottleneck. Organizational capacity to absorb accelerated output was.
Vibe coding isn't dying. It's being scoped. And the organizations that survive the disillusionment phase will be the ones that figured out where the vibes end and the engineering begins.
Sources
Research & Reports:
- 2025 GenAI Code Security Report - Veracode
- Bad Vibes: Comparing the Secure Coding Capabilities of Popular Coding Agents - Tenzai (Dec 2025)
- Army of Juniors: The AI Code Security Crisis - OX Security (2025)
- Securing Vibe Coding Tools - Palo Alto Networks Unit 42
- The Hidden Costs of AI-Generated Software - Codebridge (2026)
- Tech Debt Tsunami Building Amid AI Craze - CFO Dive / Forrester (2025)
Industry Analysis:
- Signals for 2026 - O'Reilly
- Vibe Coding, Vibe Checking, and Vibe Blogging - O'Reilly / Philip Guo
- Limitations of Vibe Coding Tools in 2026 - Builder.io
- Vibe Coding: Everything You Need to Know - Retool
- Vibe Break Chapter IV: The Lovable Inadvertence - Desplega AI
- The Uncomfortable Truth About Vibe Coding - Red Hat Developer (2026)
Commentary:
- A New Worst Coder Has Entered the Chat - Stack Overflow (Jan 2026)
- Linus Torvalds Says Vibe Coding Is Fine for Getting Started, 'Horrible Idea' for Maintenance - The Register (2025)
- Collins' Word of the Year 2025: Vibe Coding - CNN (2025)
Related Swarm Signal Coverage: