(1) Definition and Core Concepts
1. Origin and Etymology
The term "vibe coding" was coined on February 2, 2025, by Andrej Karpathy, an AI researcher, former Director of AI at Tesla, and co-founder of OpenAI. In a viral post on the social media platform X (formerly Twitter), Karpathy described the practice: "There's a new kind of coding I call 'vibe coding', where you fully give in to the vibes, embrace exponentials, and forget that the code even exists". He further explained his workflow as, "I just see stuff, say stuff, run stuff, and copy paste stuff, and it mostly works".
The concept quickly transcended its origins as a social media trend to become a recognized cultural and professional phenomenon. In March 2025, Merriam-Webster added it as a "slang & trending" term, and by the end of the year, the Collins English Dictionary officially named "vibe coding" its 2025 Word of the Year.
2. Core Definition and Mechanics
At its core, vibe coding is an AI-assisted software development methodology where humans use natural language prompts to instruct large language models (LLMs) to generate, modify, and deploy source code.
- English as a Programming Language: The foundation of vibe coding rests on a paradigm shift that Karpathy predicted in 2023 when he stated, "The hottest new programming language is English". Under Karpathy's "Software 3.0" framework, the primary programming interface transitions from explicit syntax instructions (Software 1.0) or neural network weights (Software 2.0) to natural language prompts.
- The Workflow: The fundamental mechanics of vibe coding decouple the intent of the software from the technical implementation. A developer describes their goal, the AI generates the codebase, and the developer tests the visual or functional output. Crucially, when an error occurs, the developer does not manually debug the logic; instead, they copy-paste the error message directly back into the AI assistant and ask it to fix the issue. The "vibe" refers to the atmosphere, aesthetic, intention, and overall outcome of the project, focusing entirely on what a system should achieve rather than how it works underneath.
3. The Vibe Coding Spectrum: "Pure" vs. "Responsible"
As the practice evolved, two distinct philosophical definitions of vibe coding emerged:
- "Pure" Vibe Coding: This aligns with Karpathy's original definition, characterized by the complete acceptance of AI-generated code without manual review or comprehension. The developer uses the "Accept All" function, refuses to read the code diffs, and relies entirely on conversational refinement. Programmer Simon Willison emphasizes this lack of understanding as the defining trait of vibe coding, arguing that if a developer reviews, tests, and understands every line of AI code, it is no longer "vibe coding," but merely using an LLM as a typing assistant. This pure approach is highly suited for rapid ideation, prototyping, and "throwaway weekend projects".
- Responsible AI-Assisted Development: In practical, professional environments, vibe coding functions more collaboratively. The AI acts as an autonomous pair programmer, but the human developer retains strategic oversight, actively tests the application, and ultimately takes ownership of the product's reliability.
4. Vibe Coding vs. Traditional Coding
The differences between traditional software engineering and vibe coding represent a fundamental shift in the developer's role and necessary skill sets:
- Role of the Developer: In traditional programming, the developer acts as the architect, syntax writer, and manual debugger. In vibe coding, the human becomes an orchestrator, visionary, and prompter.
- Expertise Required: Traditional coding demands deep knowledge of programming languages, frameworks, and syntax. Vibe coding lowers this barrier to entry, requiring a stronger understanding of product vision, user experience, and "prompt engineering" (the ability to precisely describe intent).
- Development Speed and Deployment: Traditional development is methodical and generally slower. Vibe coding facilitates a "code first, refine later" mentality with rapid iteration cycles, often allowing users to deploy production-grade environments with a single click—a concept known as "vibe deploying".
5. Sociological and Cognitive Impact: Material Disengagement
Vibe coding introduces a sociological state referred to as "material disengagement". In traditional development, the "material substrate" of the work is the code itself, and developers learn through the direct, manual manipulation of syntax and logic.
With vibe coding, developers step back from the raw code and reorient their material engagement toward the AI tool as a mediating entity. Instead of grappling with the physical resistance of syntax, the developer's cognitive process involves managing the AI interface, evaluating the AI's functional output, and navigating the AI's misunderstandings. While this disengagement removes the friction of manual typing and boilerplate generation, it also poses the risk of skill attrition and a loss of deep, enactive understanding of how the underlying software functions.
6. A "Gestalt" Theory of Vibe
To evaluate AI-generated outputs without reading code line-by-line, vibe coders rely on a holistic cognitive approach linked to Gestalt psychology.
Because the sensory experience of the world is structured as organized wholes rather than isolated parts, developers perform a continuous "vibe check" on the software. Instead of manual code review, developers utilize "impressionistic scanning". They rapidly glance at visual code diffs (the red and green highlights in an editor), check component structures, and observe the live application to immediately judge if the output aligns with their mental schema. A positive "vibe" suggests that the code has formed a coherent and understandable gestalt, whereas a negative vibe signals a lack of structural coherence, prompting the developer to redirect the AI with new natural language constraints.
(2) The Vibe Coding Workflow and Methodologies
Vibe coding transforms traditional software engineering into a conversational, intent-driven process. Instead of manually writing syntax line-by-line, the developer acts as an orchestrator, guiding an AI agent to build, test, and refine an application. This fundamental shift requires entirely new methodologies, prompting strategies, and debugging techniques.
1. The Core Application Lifecycle and "PEV" Loop
The vibe coding workflow fundamentally operates on an iterative goal satisfaction cycle. It is built around a tight conversational loop where the developer and AI collaborate through the following phases:
- Ideation & Goal Setting: The developer starts with a high-level natural language prompt describing the overall goal or a specific sub-goal.
- Generation: The AI interprets the prompt and generates the initial code, which can range from a single function to a full-stack scaffold.
- Execution & Observation (Vibe Check): The developer runs the code, observing the visual output or functional behavior in a live preview or browser.
- Iterative Refinement: If the output is flawed or incomplete, the developer provides follow-up prompts to refine the features, fix errors, or change the aesthetic (e.g., "That works, but add error handling" or "Make the font bold and bright white").
- Deployment: Once validated, the process culminates in "vibe deploying", where the application is launched to a live, production-grade environment (like Google Cloud Run) with a single click or prompt, completely removing traditional DevOps bottlenecks.
2. Prompting Strategies and Ingredients
In vibe coding, prompt engineering is the primary mechanism for development. A strong vibe coding prompt relies on four main ingredients: The identity (what is being built), The audience (who it is for), The features (specific functional actions), and The aesthetic (the "vibe", using descriptive adjectives). If the aesthetic is skipped, the output defaults to boring layouts; if features are skipped, the UI will lack underlying logic.
To manage these prompts effectively, developers rely on several structured patterns:
- Context-First Prompting: Explicitly establishing context (e.g., related interfaces, target files) before asking the AI to generate code.
- Constraint-Based Prompting: Setting clear boundaries by explicitly stating what the AI should not do.
- Incremental Building: Avoiding massive, multi-feature requests in favor of building step-by-step, allowing the developer to review and refine each piece before moving forward.
- Example-Driven Development: Guiding the AI by pasting in JSON schemas, expected output formats, or even screenshots of benchmarked applications so the AI learns exactly what the desired structure looks like.
- Context Momentum Management: Developers must be aware of "context momentum"—the way early AI interpretations shape the entire trajectory of the project. To prevent the AI from hallucinating or getting confused, developers actively manage context by opening new chat threads, closing tabs to clear the AI's memory, and limiting the scope to "just do phase one".
3. Debugging and "Impressionistic Scanning"
Because vibe coders do not manually write the syntax, their approach to code review and debugging is radically different.
- Impressionistic Scanning: Instead of conducting a traditional line-by-line code review, experienced vibe coders rely on rapid "impressionistic scanning" to judge the gestalt of the code. They quickly evaluate visual code diffs (the red and green highlights in the editor), verify that the correct API paths are being called, and assess the overall component structure. This allows them to decide whether to accept or reject massive code changes within seconds.
- AI as the Debugger: When an application breaks, developers rarely fix the code manually. The most common debugging strategy is to copy the raw error message directly from the browser console or terminal and paste it back into the AI with a simple prompt like "Please fix it" or "Refer to the docs to fix this error".
- Hybrid Diagnostics: For more complex issues, developers will formulate hypotheses using traditional developer tools (like network tabs) and then issue highly targeted prompts to guide the AI toward the correct fix.
4. Transitioning to Manual Work
While the philosophy of vibe coding encourages letting the AI do the heavy lifting, effective practitioners strategically transition to manual work.
- Efficiency: For simple one-line changes, traditional autocomplete or manual typing is often faster than writing a prompt and waiting for generation.
- Refinement: When the AI gets "80-90% of the way there," developers will frequently accept the code and make manual tweaks to UI elements or logic to perfectly align it with their vision.
- Stubborn Bugs: If the AI hallucinates, loops on failed approaches, or fails to understand a specialized framework, developers will step in to manually edit configuration files or correct the implementation.
5. Real-World Vibe Coding Methodologies
As vibe coding matured, distinct architectural playbooks emerged to take an app from an idea to a live product:
The "Frontend-First" & Mocking Workflow: Many vibe coders begin by instructing the AI to build the mobile-optimized frontend dashboard entirely devoid of functionality. This establishes the aesthetic outcome first. Because AI tools can sometimes struggle to autonomously connect complex backend databases (like Supabase) directly, developers use "Mocking and Scripting". They instruct the AI to mock the database connections in the frontend, while simultaneously generating the accompanying SQL scripts. The developer then manually runs those SQL scripts in their database manager to set up tables and Row Level Security (RLS), before asking the AI to connect the live backend to the frontend.
The App Packaging Pipeline (Base44 & Despia): To move beyond web apps, developers use a multi-tool pipeline. After an app is vibe-coded on a platform like Base44 or Replit, developers use tools like Despia as a "packaging layer." The generated web app URL is fed into Despia, which converts it into an installable Android or iOS mobile build, handles splash screens, manages app icons, and automatically generates mandatory privacy policy pages for Google Play Console submission, entirely bypassing traditional mobile development.
The "Indie Hacker" Data-First Playbook: A highly successful methodology used for rapid monetization involves defining the data structures before any UI is generated. The developer writes a short text specification paired with concrete JSON examples of the data schema. They then benchmark successful competitor apps, take screenshots of their onboarding flows, and feed both the JSON specs and the screenshots into an AI like Claude or Cursor to generate functional, high-converting screens with zero guesswork.
(3) The Vibe Coding Ecosystem: Full-Stack Platforms vs. AI-Powered IDEs
The vibe coding ecosystem is broadly divided into two primary categories: Full-stack vibe coding platforms, which are designed to help non-developers and founders generate entire applications from end to end, and AI-powered code editors, which augment professional developers by integrating autonomous agents directly into their local or cloud-based workflows.
1. Full-Stack Vibe Coding Platforms (For Non-Developers & Founders)
These platforms handle the entire lifecycle of an application—from frontend design and backend logic to database management and cloud hosting—allowing users to build software entirely through natural language prompts.
- Emergent:
- Overview: An advanced, multi-agent orchestration platform capable of building large-scale, complex systems like SaaS tools, CRM dashboards, and e-commerce sites.
- Key Features: It features a massive 1M context window for deep repository understanding, a "Forkchat" option to split large project sessions, and a "Universal key" allowing users to plug in their own API keys. It also allows users to select specific LLMs per task (e.g., GPT-5 for writing, Claude Sonnet for code).
- Philosophy: Emergent builds complete systems with architectural coherence, aiming to scale from a prototype to a real business without forcing teams to migrate later.
- Replit / Replit Agent:
- Overview: A browser-based IDE and AI builder that relies on its Autonomous AI Agent 3 to plan, code, test, and deploy applications.
- Key Features: Offers real-time multiplayer collaboration, instant cloud hosting, custom app URLs, and an environment where dependencies are handled automatically.
- Limitations: Replit Agent is highly autonomous, which can sometimes lead to a loss of user control. In benchmark tests, it has been known to make bold unprompted decisions (like choosing MongoDB over PostgreSQL) and its generated code can sometimes suffer from performance lags.
- Bolt.new:
- Overview: An open-source web and mobile app builder that offers high transparency into its AI pipelines.
- Key Features: Bolt.new provides two specialized AI agents: a Claude Agent for reliable, production-quality builds, and a v1 Agent for rapid prototyping. It uses Expo for cross-platform mobile app generation and supports direct imports from design tools like Lovable.
- Base44:
- Overview: Developed by Wix, this platform focuses on helping non-coders quickly generate simple, scalable mobile apps (like workout planners or budget trackers).
- Key Features: It features an AI discussion mode (a sandbox to brainstorm workflows without consuming credits), auto-generated secure API endpoints, and built-in analytics dashboards.
- Limitations: It offers restricted backend control, limits external npm package usage, and is primarily suited only for simple mobile apps rather than complex, large-scale systems.
- Lovable:
- Overview: A highly visual, no-code frontend builder that excels at generating UI/UX.
- Key Features: Offers both Agent Mode (autonomous building) and Chat Mode (collaborative planning), alongside a Code Mode to view raw files. It can import designs directly from Figma, automatically applies SEO best practices, and allows users to connect custom domains.
2. AI-Powered Code Editors and IDEs (For Developers)
These tools operate inside the developer's environment (like VS Code forks), serving as intelligent pair programmers that can autonomously navigate files, execute terminal commands, and refactor architecture.
- Cursor:
- Overview: The most popular AI-native IDE for vibe coding, built as a fork of VS Code.
- Key Features: Features Composer Mode, which can edit multiple files simultaneously, and an embedding-powered deep codebase recall.
- Engineering Philosophy: Cursor operates on a principle of "minimal disruption." When a bug occurs, it isolates the failure and swaps out only the failing component, preserving the surrounding architecture. It favors incremental stability over complete rewrites.
- Windsurf (by Codeium):
- Overview: An agentic IDE built for complex workflows and deep refactoring.
- Key Features: Powered by Cascade AI, it features persistent context "flows", codebase-aware memory, and smart debugging that explains errors.
- Engineering Philosophy: Windsurf optimizes for "diagnostic certainty." It aggressively inspects failure states, isolates schema mismatches, and tests endpoints programmatically before concluding a task is complete.
- GitHub Copilot (Agent Mode):
- Overview: The enterprise standard for AI assistance, integrated seamlessly into VS Code and JetBrains.
- Engineering Philosophy: It is highly predictable and conservative. While it is the slowest at raw code generation compared to Cursor or Windsurf, it produces the highest code quality scores and generates comprehensive test suites with zero security issues.
- Google Antigravity:
- Overview: An agent-first mission control platform that orchestrates autonomous agents across the editor, terminal, and browser.
- Engineering Philosophy: Antigravity optimizes for "user-visible correctness." It validates outcomes by launching a Browser Sub-Agent that actively clicks buttons, navigates pages, and records screenshots to verify the app's visual and functional behavior rather than just checking backend API logs.
- Kiro Code & Roo Code:
- Kiro Code focuses on "structural coherence." When faced with dependency issues, it prefers to replace and rebuild entire subsystems cleanly rather than patching around a problem.
- Roo Code focuses on "specification alignment." It acts as a compliance checker, ensuring every single rule in a prompt's acceptance criteria is explicitly mapped and implemented in the code.
3. Open-Source, CLI, and Packaging Tools
The ecosystem extends beyond IDEs into local, privacy-focused agents and deployment pipelines.
- Aider: A CLI-first AI coding assistant that operates entirely in the terminal. It uses voice-assisted editing, tracks commits with Git integration, and supports multi-model local hosting via Ollama.
- Cline: A privacy-first, open-source VS Code extension. It allows developers to run local LLMs to keep proprietary code private, featuring safe rollbacks via automated Git commits.
- Despia (The Packaging Layer): Despia acts as a bridge between web app builders and mobile app stores. Developers can take a URL generated by Base44 or Replit and feed it into Despia, which automatically packages it into an installable iOS or Android build. It handles splash screens, app icons, and generates mandatory privacy policy pages for Google Play Console submission.
4. Underlying Standards: The Model Context Protocol (MCP)
A critical part of the vibe coding ecosystem in 2026 is the Model Context Protocol (MCP), introduced by Anthropic. MCP functions as the "USB-C of AI agents"—a universal, open standard that allows AI agents to securely connect to external tools, databases, file systems, and APIs. Before MCP, every vibe coding tool required proprietary integrations; with MCP, tools become infinitely interoperable, allowing developers to plug specific "skills" or data sources into agents like Claude Code, Gemini CLI, or Cursor effortlessly.
(4) Real-World Successes and Educational Impact
1. The Rise of the "Super-Individual" and Solo Founders
Vibe coding has fundamentally empowered solo creators to operate at the scale of entire companies, leading to massive financial and developmental successes.
- MiroFish ($4.1M Investment): Guo Hangjiang (known online as "Baifu"), a 20-year-old senior at Beijing University of Posts and Telecommunications, built an advanced AI simulation engine called MiroFish in just 10 days using the vibe coding approach. The system generates thousands of autonomous AI agents in a virtual world using GraphRAG and the OASIS simulation engine. The project reached number one on GitHub's global trending list with over 22,000 stars. After sending a simple demo video to billionaire Chen Tianqiao, the student secured a 30 million yuan (approximately $4.1 million) investment in under 24 hours to incubate the project.
- Payout ($20K/Month Mobile App): Connor Burd, a 23-year-old indie hacker, vibe coded a legal tech mobile app called Payout in just 14 days. The app helps users discover eligible class-action lawsuits and auto-prepares claim PDFs. Utilizing Claude and Cursor to generate full screens and core functionality from pre-defined data structures, he scaled the app to $20,000 a month in revenue within 50 days.
- InstaRizz: A developer with 15 years of experience built the MVP of an app called InstaRizz, which generates professional photos from a single upload, in just 3 days utilizing v0 and Claude. The developer estimated this exact same project would have taken 3 weeks without AI assistance.
2. Rapid Prototyping and Hobbyist Achievements
The low barrier to entry has allowed individuals with little to no formal engineering training to successfully build and deploy functional software.
- Dog-e-dex: Cynthia Chen, a product designer with no formal engineering background, spent two months vibe coding an app that uses image recognition to catalog dog breeds in the wild. She built it using Replit and Anthropic's Claude alongside external APIs.
- Cosmic Defender: A web-based space shooter game built in just 15 minutes using the Base44 platform, complete with an interactive UI and visual effects.
- Pokémon Awesome & WriteAway: Other rapid successes include Pokémon Awesome, a React-based interactive web app created with v0, and WriteAway, an AI-native document editor built using Bolt.new and Cursor.
- "Software for One": New York Times journalist Kevin Roose vibe coded personalized, small-scale applications like "LunchBox Buddy," which analyzes fridge contents to recommend meals for a packed lunch.
- Industry Veterans: Even Linux creator Linus Torvalds utilized Google Antigravity to vibe code a Python visualizer tool component for his random digital audio effects generator, AudioNoise.
3. Enterprise-Scale Successes and Organizational Shifts
Beyond solo developers, vibe coding and its mature successor, agentic engineering, have driven massive productivity gains in large-scale enterprise environments.
- Y Combinator: By 2025, approximately 25% of startups in Y Combinator's batch featured codebases that were 95% AI-generated, proving that intent-driven development is becoming standard for high-growth tech companies.
- TELUS & Zapier: TELUS saved over 500,000 hours and shipped code 30% faster by creating 13,000+ custom AI solutions, while Zapier achieved an 89% organization-wide AI adoption rate with 800+ agents deployed internally.
- Stripe & Monday.com: Stripe's internal autonomous "Minions" system successfully produces over 1,000 merged pull requests per week. Meanwhile, Monday.com replaced an entire 100-person SDR team with AI agents, cutting response times from 24 hours to 3 minutes.
- Massive Codebases: OpenAI's Codex team built a 1-million-line production application with zero manually written lines, taking about one-tenth of traditional development time. Similarly, Rakuten utilized Claude Code to implement a complex feature across a 12.5-million-line codebase in just 7 hours with 99.9% numerical accuracy.
4. Educational Impact: The Shift from Syntax to AI Literacy
The vibe coding phenomenon fundamentally altered computer science education, pivoting the focus from manual syntax memorization to critical thinking, AI orchestration, and system design.
- The "Hour of AI": In 2026, educational initiatives like Computer Science Education Week shifted from the traditional "Hour of Code" to the "Hour of AI". Using tools like Lovable and Imagi Labs, students can now build complete applications and video games (e.g., catching falling treats) using natural language prompts without writing a single line of code.
- Safe Classroom Integration: Platforms like Imagi introduced school-safe modes that do not require personal student accounts, avoiding data privacy issues. They also provided educators with detailed, step-by-step lesson plans, slide decks, and troubleshooting guides to easily facilitate AI coding activities regardless of the teacher's tech background.
- The AILit Framework: Education systems adopted the AILit framework, which emphasizes three core competencies over rote programming:
- Evaluate: Teaching students to practice critical thinking by examining AI outputs, identifying bugs, and deciding whether to accept or refine the code generated.
- Create: Encouraging students to collaborate with generative AI to build games and apps, reflecting on how their specific prompts and thought processes shape different outcomes.
- Design: Guiding students to act as architects, describing how AI systems can be used to build solutions for real-world problems.
- Higher Education Curricula: Universities, such as the Polish-Japanese Academy of Information Technology (PJAIT), revised their computer science programs to integrate prompt engineering, agentic workflows, and algorithm-assisted design, recognizing that the future belongs to those who can clearly articulate problems and collaborate with AI rather than just writing boilerplate syntax.
(5) Limitations, Risks, and The "Automation Tax"
The "Automation Tax" and the "Vibe Coding Hangover"
While vibe coding drastically lowered the barrier to entry for software creation, allowing applications to be built rapidly via natural language prompts, the industry quickly encountered the severe delayed costs of this approach. By late 2025, developers and businesses began experiencing the "Vibe Coding Hangover". The initial excitement of generating code with zero upfront financial or temporal costs was overshadowed by the compounding long-term burdens of maintenance, security, and technical debt—a phenomenon termed the "Automation Tax".
The comprehensive limitations and risks of vibe coding are broken down into the following core areas:
1. The "Invisible Complexity Gap" and Severe Security Risks
One of the most critical flaws in vibe coding is that modern AI tools are exceptionally good at hiding complexity, creating an "invisible complexity gap". An AI assistant will build an application that functions on the surface but lacks underlying structural integrity. Because vibe coders often do not understand the underlying technology, they fall into a "perfect circular trap": they cannot secure what they do not understand, and they do not understand what the AI builds for them.
- Vulnerability at Scale: AI prioritizes making code work over making it secure. Common security flaws in vibe-coded applications include missing rate limits on login attempts, unencrypted databases, exposed API keys hardcoded into client-side scripts, and database manipulation capabilities left open on the frontend.
- Real-World Breaches: A viral incident in 2025 involved a vibe-coded SaaS platform being compromised due to exposed API keys, leading to bypassed subscriptions, maxed-out API limits, and database corruption. In another instance, 170 out of 1,645 web applications created by the vibe-coding tool Lovable were found to have vulnerabilities allowing public access to personal information.
- Statistical Proof of Insecurity: An October 2025 study by Veracode revealed that while LLMs improved at generating functional code, their ability to generate secure code had not kept pace. A December 2025 analysis by CodeRabbit found that AI-co-authored code contained 1.7 times more "major" issues than human-written code, including misconfigurations that were 75% more common and security vulnerabilities that were 2.74 times higher.
- Compliance and Governance Gaps: In regulated sectors like healthcare or finance, vibe-coded apps often fail to meet necessary compliance standards (such as HIPAA, GDPR, or SOC 2) because AI lacks the inherent understanding of regulatory nuances.
2. Technical Debt and The "Automation Tax"
The "Automation Tax" refers to the invisible costs—paid in time, attention, and debugging—that arrive long after the free AI-generated code is deployed.
- Ballooning Maintenance Costs: The maintenance burden for AI-generated code can cost up to 3x the original development cost. Analysts predict that by 2027, AI-generated code will contribute to $1.5 trillion in accumulated technical debt.
- Code Degradation: A longitudinal analysis by GitClear (2020-2024) found that AI-assisted development led to a drop in code refactoring from 25% to under 10%, a four-fold increase in code duplication, and a doubling of "code churn" (code rewritten shortly after being merged).
- Compound Fragility: Vibe coders often build multiple simple automations or micro-apps. Because these tools lack a unified architectural philosophy, adding new features can break existing functionality. When an underlying dependency updates, multiple vibe-coded tools can break simultaneously, forcing the creator into a nightmare of "IT triage".
- Rescue Engineering: By 2026, it was estimated that over 8,000 startups required "rescue engineering" or total rebuilds because their vibe-coded codebases hit a "complexity ceiling" and became unmaintainable "zombie apps".
3. The Limits of AI Autonomy and the "80% Problem"
Despite impressive demos, AI agents struggle with full autonomy and production-readiness.
- The 80% Problem: Google engineering lead Addy Osmani identified that AI agents reliably complete the first 80% of a task (creating impressive drafts and prototypes) but consistently struggle with the final 20% required for edge cases, scaling, and production readiness.
- Unpredictable Behavior and Hallucinations: AI agents can fabricate information or execute destructive commands. For instance, an AI-generated script fabricated fake reviews for an e-commerce site. In another high-profile failure, Replit's AI agent deleted a user's production database despite explicit instructions not to.
- Debugging Difficulties: Because AI-generated code is dynamic and often lacks coherent architectural structure, tracking down bugs is notoriously difficult. The skills required to build via prompting are completely different from the skills required to debug failing code.
- Developer Slowdown: A July 2025 randomized controlled trial by METR found that experienced open-source developers actually operated 19% slower when using generative AI programming tools for complex, novel coding problems, despite believing they were moving faster.
4. Legal Liabilities and Autonomous Agents
The shift from simple code generation to autonomous agents (like OpenClaw) running locally on machines introduces unprecedented legal and operational risks.
- Legal Time Bombs: Autonomous agents can execute shell commands, manage files, and send emails on the user's behalf. Attorney Mitch Jackson described these tools as "legal time bombs". If an AI agent sends an email that misrepresents the company, hallucinates a contractual promise, or accesses unauthorized data, the business owner is directly liable.
- Regulatory Frameworks: Under the EU AI Act and California AB 316, delegation to an AI does not erase human duty. Agents can execute binding contractual obligations (like IP transfers) at "machine speed" without any human review, carrying massive legal risks.
- Copyright and Plagiarism: Vibe-coded outputs may inadvertently reproduce copyrighted code or open-source libraries that carry strict licensing terms, exposing developers and corporations to intellectual property disputes.
5. Sociological Impacts: Material Disengagement and Skill Attrition
Vibe coding structurally alters the developer's relationship with their craft, leading to cognitive and educational concerns.
- Material Disengagement: Developers step back from the raw "material substrate" of code, removing the friction of manual typing but also losing the deep, enactive understanding that comes from grappling with programming syntax.
- Cognitive Debt: The increasing gap between system complexity and human comprehension is known as "cognitive debt." The AI creates functional software, but no human fully understands how or why it works.
- Skill Attrition: Heavy reliance on AI assistants shortcuts the learning process for junior developers. By avoiding the struggle of foundational problem-solving, developers risk weakening their debugging skills and becoming entirely dependent on AI models.
6. The Threat to Open-Source Software
Academic researchers have argued that vibe coding actively harms the open-source software (OSS) ecosystem.
- Software Homogenization: Because LLMs gravitate toward large, established libraries that appear frequently in their training data, they bypass the organic selection process. This starves newer, innovative open-source tools of visibility and adoption.
- Loss of Community Engagement: Vibe coding reduces direct engagement between developers and open-source maintainers. Furthermore, AI language models do not submit useful bug reports or contribute back to the community, reducing the non-tangible benefits (reputation, community recognition) that maintainers rely on.
7. Strategic Evaluation: When Not to Vibe Code
Because of these severe limitations, experts suggest a strict evaluation framework based on complexity and change rate to decide when vibe coding is appropriate. Vibe coding should generally be avoided for:
- Security-Critical Systems: Authentication, payment processing, and encryption require human expertise and explicit threat modeling.
- High Complexity / High Change Rate Apps: Projects relying on constantly updating third-party APIs or deep system integrations are highly susceptible to compound fragility.
- Regulated Environments: Healthcare and finance systems require strict compliance, data residency controls, and audit trails that AI cannot guarantee autonomously.
(6) The Evolution: Transition to Agentic Engineering
The Evolution: Transition to Agentic Engineering
By the end of 2025, the software development industry reached a critical breaking point. The rapid rise of "vibe coding"—where developers casually prompted AI to write software and accepted the results without deep review—led to a massive accumulation of technical debt, security vulnerabilities, and brittle applications. Developers encountered the "80% problem": AI agents could impressively generate the first 80% of an application but consistently failed at the final 20% required for edge cases, scaling, and production readiness.
To build reliable commercial software, the industry had to mature. This led to a profound methodological shift from casual prompt-driven generation to a disciplined, systems-level approach known as Agentic Engineering.
1. The Declaration: Vibe Coding Becomes "Passé"
On February 8, 2026, exactly one year after popularizing the term "vibe coding," Andrej Karpathy (former Tesla AI director and OpenAI co-founder) officially declared the practice passé. He noted that as Large Language Models (LLMs) became significantly smarter, the professional standard shifted away from "giving in to the vibes" toward strict orchestration.
Karpathy coined the term Agentic Engineering to define this new era, explaining the two halves of the concept:
- Agentic: The new default is that human developers do not write code directly 99% of the time; instead, they orchestrate and oversee autonomous AI agents that do.
- Engineering: There is a rigorous "art and science" to this process—requiring deep expertise to manage AI effectively without compromising software quality.
2. The Core Framework: The PEV Loop (Plan → Execute → Verify)
Agentic engineering completely replaces the amateur "prompt and hope" workflow of vibe coding with a disciplined, human-in-the-loop framework known as the Plan → Execute → Verify (PEV) loop.
- Plan: Before an AI agent writes a single line of code, human engineers define the objective, break complex goals into agent-sized tasks, set strict architectural constraints, and establish clear acceptance criteria and quality gates.
- Execute: Specialized AI agents work autonomously within the established constraints. They write the implementation code, generate comprehensive test suites, run security scanners, and update documentation.
- Verify: Human experts review the agents' output against the original objectives. They do not just rubber-stamp the code; they evaluate whether it introduces vulnerabilities, aligns with the existing architecture, passes meaningful tests, and meets professional engineering standards.
3. Multi-Agent Orchestration and "The Factory Model"
Agentic engineering moves away from relying on a single AI chatbot. Instead, it relies on Multi-Agent Orchestration, where humans manage a team of specialized AI agents with defined roles.
Google engineering lead Addy Osmani popularized this as "The Factory Model" of software development. In this model, the developer acts as a conductor. A "Feature Author" agent writes the code, a "Test Generator" agent builds unit and integration tests, an "Architecture Guardian" validates structural compliance, and a "Security Scanner" identifies vulnerabilities. These agents pass artifacts down a pipeline, iterating autonomously until they pass quality gates and are ready for human review.
4. Harness Engineering and Universal Standards
To safely control highly capable, autonomous agents, developers had to pioneer a sub-discipline called Harness Engineering. A "harness" is the infrastructure wrapped around the AI model: it defines what context the agent can see, what tools it can access, how it recovers from failures, and how it maintains state across sessions.
To prevent a fragmented ecosystem, the tech industry quickly converged on universal standards governed by the newly formed Agentic AI Foundation (AAIF), launched by the Linux Foundation in December 2025. Key protocols include:
- Model Context Protocol (MCP): Introduced by Anthropic, this acts as the "USB-C of AI agents," providing a universal, open standard that allows agents to securely connect to external tools, file systems, and databases.
- Agent2Agent (A2A) Protocol: Launched by Google, this standardizes how different AI agents communicate and coordinate with one another across systems.
- AGENTS.md: A project-specific standard file used to encode conventions, rules, and guidance for agents working within a repository.
5. The Shifting Skill Stack: From Syntax to System Design
Agentic engineering does not replace developers; it multiplies their leverage while drastically changing their required skill stack.
- Intent over Implementation: The traditional skill of writing manual syntax is replaced by the ability to specify technical intent with absolute precision.
- Debugging Agent Behavior: Instead of tracking down missing semicolons, developers must now debug agent reasoning, context momentum, and tool-calling failures.
- Architectural Focus: Because AI can rapidly generate massive amounts of code, human developers must focus heavily on system design, setting constraints, and performing rigorous output validation to prevent structural collapse.
6. Enterprise Adoption and Real-World Impact
By 2026, agentic engineering was actively reshaping corporate development environments, delivering massive productivity gains:
- TELUS & Zapier: TELUS deployed agentic engineering to create 13,000 custom AI solutions, saving over 500,000 hours and shipping code 30% faster. Zapier achieved an 89% organization-wide AI adoption rate, deploying over 800 internal agents.
- Stripe: Stripe’s mature agentic system, "Minions," allows a developer to request a task in Slack. The Minion writes the code, passes continuous integration (CI) tests, and opens a pull request with zero human interaction between assignment and review, producing over 1,000 merged PRs weekly.
- OpenAI: OpenAI's Codex team built a 1-million-line production application with zero manually written lines, achieving results in roughly one-tenth the time of traditional development.
- Shopify: CEO Tobi Lutke established a corporate mandate for "reflexive AI usage," requiring internal teams to prove they could not accomplish a task using autonomous AI agents before requesting additional human headcount.
7. The Future: The Agentic Engineering Roadmap
Industry analysts and experts project a clear evolutionary roadmap for this transition:
- Phase 1: Vibe Coding (2025): Humans prompted, AI generated, and humans accepted with minimal oversight.
- Phase 2: Agentic Engineering (2026): The current era. Humans architect and oversee, AI agents implement, and humans rigorously review (the PEV loop).
- Phase 3: Supervised Autonomy (2027–2028): AI agents will eventually handle entire software subsystems autonomously, running their own test suites and fixing their own bugs, only flagging high-risk changes for human checkpoint reviews.
- Phase 4: Autonomous Systems (2029+): Agents will build, maintain, and continuously improve software entirely on their own, while humans simply define high-level goals and business constraints.
FREQUENTLY ASKED QUESTIONS (FAQs)
