The origin story
In February 2025, Andrej Karpathy coined the term "vibe coding" in a post on X. He described it as a new way to write software where you "fully give in to the vibes, embrace exponentials, and forget that the code even exists."
The post went viral. Within weeks, "vibe coding" had its own Wikipedia page. Thousands of developers started building apps without reading the code their AI assistants generated. Non-technical founders launched startups built entirely on AI-generated codebases. The promise was seductive: anyone can build software now.
But buried in that same original post was a caveat most people ignored: "Sometimes the LLMs can't fix a bug so I just work around it or ask for random changes until it goes away." Karpathy himself called it suitable for "throwaway weekend projects." Not production systems. Not startups. Not software that handles user data.
One year later, the data is in. And it paints a very different picture than the hype suggests.
What the data says
Let's start with the largest study of its kind. CodeRabbit analyzed 470 GitHub pull requests, comparing AI-generated code against human-written code across multiple dimensions. The results were not close.
Where AI makes more mistakes: Code quality.
Takeaway: AI creates 1.7x more issues than human code.
AI-generated code creates 1.7x more issues overall than human-written code. But the breakdown matters. Logic errors (the kind that cause real production bugs) were 1.75x more frequent. Concurrency issues, which are notoriously hard to debug, were 2.29x more common. These are not cosmetic problems. These are the bugs that wake you up at 3 AM.

2025 was the year the internet broke: Studies show increased incidents due to AI
Pull requests per author increased 20% YoY, while incidents per PR rose 23.5% and change failure rates rose around 30%.
coderabbit.ai
The code quality crisis
GitClear conducted the most comprehensive analysis of AI's impact on code quality, studying 211 million changed lines of code from 2020 to 2024. The trends are alarming.
The maintainability collapse.
211 million lines of code, 2020 vs 2024
Takeaway: More output, less understanding, worse code.
The story these numbers tell is clear: developers are generating more code than ever, but understanding less of it. Refactored code (code that has been reorganized and improved) collapsed from 24.1% to just 9.5%. Meanwhile, duplicated code surged. AI tools optimize for output, not for maintainability.
As Simon Willison, creator of Datasette and a leading voice on responsible AI usage, put it:
Willison draws a critical distinction: "If an LLM wrote every line of your code, but you've reviewed, tested, and understood it all, that's not vibe coding. That's using an LLM as a typing assistant." The problem is not AI assistance. The problem is blind trust.
The security nightmare
If the quality data is concerning, the security data is terrifying. A study by SoftwareSeni found that AI-generated code contains 2.74x more security vulnerabilities than human-written code. And these are not theoretical risks.
Where AI makes more mistakes: Security vulnerabilities.
Takeaway: AI makes dangerous security mistakes.
45%
AI code with flaws
Veracode 2025
86%
Had XSS errors
AI-generated cases
74
CVEs confirmed
Georgia Tech

Securing Vibe Coding Tools: Scaling Productivity Without Scaling Risk
88% of AI-generated logging code failed to sanitize inputs. SQL injection was still observed in 20% of samples.
unit42.paloaltonetworks.com
Georgia Tech's Vibe Security Radar project has been tracking CVEs directly caused by AI-generated code since May 2025. Their findings: 74 confirmed CVEs attributed to AI coding tools, with the count accelerating: 6 in January 2026, 15 in February, 35 in March. The trend line is going in the wrong direction.
The productivity paradox
Perhaps the most surprising finding comes from METR (Model Evaluation & Threat Research), which ran a randomized controlled trial (the gold standard in research) on 16 experienced open-source developers completing 246 tasks in repositories they had an average of 5 years of experience with.
The productivity paradox.
Randomized controlled trial, 16 experienced developers, 246 tasks
Devs predicted
24%
faster
Devs believed after
20%
faster
Actual result
19%
SLOWER
Takeaway: 39-point gap between perception and reality.
Read that again. Developers believed AI made them 20% faster. In reality, they were 19% slower. That's a 39-percentage-point perception gap. The study found that developers accepted fewer than 44% of AI generations, meaning most of the time was spent reviewing, testing, and ultimately rejecting suggestions.
The explanation is straightforward: AI excels at generating plausible-looking code for small, well-defined tasks. But experienced developers working in large, mature codebases are doing something fundamentally different. They are navigating years of architectural decisions, implicit conventions, and domain-specific logic that no LLM has seen.
Real-world failures
The statistics become more concrete when you look at the actual incidents. Here are documented cases from the past year where vibe-coded applications failed in production:

Social networking app built entirely via vibe coding. Misconfigured database exposed 1.5 million authentication tokens and 35,000 email addresses.
Towards Data Science ↗
CVE-2025-48757: missing Row Level Security on Supabase tables exposed 170+ production applications to full database access.
CVE Database ↗
Broken access controls let unauthorized users bypass auth and register for private applications, exposing every app on the platform.
Accorian Security ↗
CVE-2025-54135: remote code execution vulnerability in Cursor itself, with no user interaction required.
Security Advisory ↗
See all documented vibe coding failures
34 incidents tracked with authoritative sources
These are not edge cases. They are the predictable outcome of shipping code that nobody reviewed. And as vibe coding adoption accelerates, the incident rate is accelerating with it.
Even Karpathy agrees
Perhaps the most telling signal comes from the inventor of the term himself. When Karpathy built his latest project, Nanochat (a minimal ChatGPT-like interface), he did not vibe code it.

Inventor of Vibe Coding Admits He Hand-Coded His New Project
Even Andrej Karpathy, who coined "vibe coding," built Nanochat by hand: "It's basically entirely hand-written."
futurism.com
When the stakes were real (when he was building something meant to last, not a weekend prototype) he chose to write the code himself. That tells you everything about where vibe coding sits in the hierarchy of software development: it is a prototyping tool, not a replacement for engineering.
What AI is actually good at
None of this means AI coding tools are useless. Quite the opposite. The data shows clear areas where AI excels:
Boilerplate
Test files, configs, CRUD
Security patches
20x faster per vuln fix
Code translation
Cross-language, explanation
Migrations
Framework bumps, dep updates
The pattern is clear: AI tools excel at well-defined, bounded tasks with verifiable outputs. They struggle with ambiguity, architectural decisions, cross-cutting concerns, and anything requiring understanding of business context. In other words, exactly the things that make software engineering hard.
The job market reality
If vibe coding was truly replacing developers, we would see it in the labor market. We don't.
15%
Job growth
2024-2034, BLS
$133k
Median salary
Developers, 2024
4.4M
US engineers
Growing, 2026
Software Developers: Occupational Outlook Handbook
Employment projected to grow 15% from 2024 to 2034, much faster than average. About 129,200 openings per year.
bls.gov
Software Engineers in US 2026: 4.4M Devs & Trends
The US faces a projected shortfall of 1.2 million developers by 2027. The ratio of CS graduates to open positions is 1:3.5.
boundev.com
The real concern is not that developers will be replaced, but that the junior developer pipeline is being disrupted. Entry-level tech hiring has decreased 25% year-over-year, and employment for developers aged 22 to 25 is down nearly 20% from its 2022 peak. Companies are using AI to handle tasks that used to train junior developers, without considering how the next generation of senior engineers will develop the judgment and debugging instincts that make them effective.

AI vs Gen Z: How AI has changed the career pathway for junior developers
Debugging someone else's code teaches you how systems fail. Writing boilerplate teaches structure. Reviewing PRs teaches you how others think.
stackoverflow.blog
The verdict
The evidence is overwhelming. Vibe coding (the practice of generating code you never read, never review, and never understand) produces software that is buggier, less secure, harder to maintain, and paradoxically slower to develop. Every major study published in the past year points in the same direction.
This does not mean AI coding tools are bad. It means they are tools, not replacements. The most effective use of AI in software development looks nothing like vibe coding. It looks like an experienced engineer using AI to generate first drafts, then reviewing every line. Testing rigorously. Understanding the architecture. Making deliberate decisions.
Simon Willison calls this "vibe engineering." Whatever you call it, it requires the one thing vibe coding explicitly throws away: understanding your code.
The developers who thrive in this era will not be the ones who outsource their thinking to an LLM. They will be the ones who use AI as a force multiplier on top of real engineering skills. If you are preparing for a career in software, the fundamentals matter more than ever: data structures, algorithms, system design, debugging, and the ability to reason about complex systems.
Vibe coding is a party trick. Software engineering is a discipline. The data says one of them has a future. It is not the party trick.
Sources

AI vs human code gen report: AI code creates 1.7x more issues
We analyzed 470 open-source GitHub pull requests, using CodeRabbit's structured issue taxonomy and found that AI generated code creates 1.7x more issues.
coderabbit.ai

Securing Vibe Coding Tools: Scaling Productivity Without Scaling Risk
AI-generated code looks flawless until it isn't. Unit 42 breaks down how to expose these invisible flaws before they turn into your next breach.
unit42.paloaltonetworks.com
Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity
A randomized controlled trial on 16 experienced developers completing 246 tasks. AI tools made them 19% slower, despite believing they were 24% faster.
metr.org

Inventor of Vibe Coding Admits He Hand-Coded His New Project
Even former OpenAI exec Andrej Karpathy, who coined the term "vibe coding," has seemingly fallen out of love with his own creation.
futurism.com

AI vs Gen Z: How AI has changed the career pathway for junior developers
For promising Gen Z students, a career as a software developer seemed like the golden ticket to career stability and success. But in the age of AI, the promise is gone.
stackoverflow.blog
AI Copilot Code Quality: 2025 Data Suggests 4x Growth in Code Clones
Research examining how AI assistants influence code quality across 211 million changed lines of code from major technology companies.
gitclear.com
Additional sources
CodeRabbit: Why 2025 Was The Year The Internet Kept Breaking
SoftwareSeni: AI-Generated Code Security Risks (2.74x)
Veracode / Infosecurity Magazine: AI Code Vulnerabilities
US Bureau of Labor Statistics: Software Developer Outlook
Towards Data Science: Vibe Coding and the Security Debt Crisis
Accorian: Security Impact of Vibe Coding (Deep Dive)
Simon Willison: Not All AI-Assisted Programming Is Vibe Coding

