<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Developer Workflow on RockB</title><link>https://baeseokjae.github.io/tags/developer-workflow/</link><description>Recent content in Developer Workflow on RockB</description><image><title>RockB</title><url>https://baeseokjae.github.io/images/og-default.png</url><link>https://baeseokjae.github.io/images/og-default.png</link></image><generator>Hugo</generator><language>en-us</language><lastBuildDate>Sat, 16 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://baeseokjae.github.io/tags/developer-workflow/index.xml" rel="self" type="application/rss+xml"/><item><title>Agentic Engineering: The Developer Guide Beyond Vibe Coding</title><link>https://baeseokjae.github.io/posts/agentic-engineering-guide-2026/</link><pubDate>Sat, 16 May 2026 00:00:00 +0000</pubDate><guid>https://baeseokjae.github.io/posts/agentic-engineering-guide-2026/</guid><description>92% of developers now use vibe coding, yet 63% waste more time debugging AI-generated code than writing it. This guide covers the practices, tools, and workflow of agentic engineering for production-ready development.</description><content:encoded><![CDATA[<p>By early 2026, 92% of US-based developers had adopted vibe coding in some form. The appeal is obvious: describe what you want in plain language, let the AI generate the code, and ship faster than ever before. But a counter-trend has emerged just as quickly. Developers who pushed vibe coding into production-grade systems discovered that speed without oversight creates a new category of technical debt — one that is especially hard to unwind because there is no specification to return to. Agentic engineering is the structured answer: a deliberate workflow that keeps human engineers in command of AI agents rather than surrendering judgment to them. This guide covers everything you need to make the shift — the principles, the practices, the tools, and the repeatable workflow that separates prototypes from production systems.</p>
<h2 id="agentic-engineering-vs-vibe-coding-the-critical-distinction">Agentic Engineering vs Vibe Coding: The Critical Distinction</h2>
<p>The 92% adoption rate for vibe coding tells only half the story — it does not tell you what happened after the code shipped. Vibe coding is prompt-driven, intuition-led development where a developer describes desired behavior to an AI model and accepts the output without deep review. It optimizes for momentum: you stay in flow, the AI generates files, you run the app, and if it works visually, you move on. Agentic engineering inverts this dynamic. The human engineer defines architecture, constraints, and validation criteria up front, then orchestrates AI agents to execute specific tasks within those boundaries. Every agent output passes through a review checkpoint before the next step begins. The distinction is not which AI model you use or how fast you type — it is who holds engineering judgment. In vibe coding, the AI holds it by default. In agentic engineering, the human retains it deliberately and uses AI to accelerate execution within well-defined guardrails. This difference is invisible at the prototype stage and becomes the deciding factor in every production deployment, every security audit, and every handoff between team members.</p>
<h2 id="why-vibe-coding-fails-at-scale-the-63-debugging-problem">Why Vibe Coding Fails at Scale: The 63% Debugging Problem</h2>
<p>Sixty-three percent of developers report spending more time debugging AI-generated code than it would have taken to write the same code themselves — a productivity paradox that exposes the hidden cost of the vibe coding model. The problem is structural, not accidental. When you accept AI output without writing a specification first, you have no ground truth to debug against. A function may pass a visual smoke test and still contain logic errors, race conditions, or edge-case failures that only surface under production load. Security compounds the issue dramatically: 40 to 62% of AI-generated code contains security vulnerabilities, and AI-written code produces flaws at measurably higher rates than human-written code. Trust has followed accordingly — developer trust in AI accuracy dropped from over 70% in 2023 to just 29% in 2026. Developers are using AI more and trusting it less simultaneously, which is a coherent response to observed reality. The debugging problem is worst when the AI generates large, interconnected diffs without intermediate checkpoints. A 500-line feature implementation that looks right in aggregate can embed a broken assumption on line 47 that invalidates everything downstream. Agentic engineering addresses this by making specifications explicit before generation and keeping changes small, reviewable, and verifiable at every increment.</p>
<h2 id="context-management-the-foundation-of-reliable-agent-output">Context Management: The Foundation of Reliable Agent Output</h2>
<p>Every reliable agentic workflow begins with persistent project context — 80% of developers using AI agents report that cold-starting a new session without providing prior context is the single largest source of irrelevant or incorrect AI output. Context management is the practice of capturing and maintaining the information an AI agent needs to produce consistent, architecture-aligned code across sessions. The primary mechanism for this in 2026 is a persistent context file: CLAUDE.md for Claude Code users, AGENTS.md for tools that follow the OpenAI agents convention. These files contain the project&rsquo;s tech stack, directory structure, naming conventions, architectural constraints, external API dependencies, security requirements, and any decisions made in previous sessions that the agent should treat as settled. A well-maintained CLAUDE.md file removes the need to re-explain project fundamentals at the start of every session. It also constrains the AI&rsquo;s solution space — when the file specifies that the project uses PostgreSQL with a specific schema design pattern, the agent stops suggesting alternative database approaches that would break existing code. Effective context files are living documents updated after every significant agent session. When the agent makes an architectural decision, it gets captured. When a pattern is established, it gets documented. The overhead is small; the payback is consistent, coherent output from the very first token of each new session.</p>
<h2 id="test-driven-agentic-development-let-agents-verify-their-own-work">Test-Driven Agentic Development: Let Agents Verify Their Own Work</h2>
<p>Test-driven development pairs naturally with agentic workflows because it gives the agent a verifiable success criterion before a single line of implementation exists. The sequence matters: write the failing test first, hand the test to the agent as the task definition, and let the agent generate implementation code until the test passes. This is fundamentally different from asking an agent to implement a feature and then hoping it works — the agent has an objective, machine-checkable feedback loop that does not require human review of every intermediate state. Developers using test-first agentic workflows report that agents consistently produce tighter, more correct implementations when given tests to satisfy rather than open-ended feature descriptions. The reason is straightforward: a failing test is a specification. It defines inputs, expected outputs, and edge cases precisely, leaving less room for the agent to make assumption errors. Teams that have adopted test-driven agentic development as a standard practice report significantly fewer regression bugs per feature and meaningfully shorter review cycles, because the code arrives pre-verified against explicit requirements rather than a verbal description. The workflow also makes refactoring safer — when you ask an agent to restructure existing code, a passing test suite is the proof that the refactor did not break behavior. Agents can run tests autonomously between each change, catching regressions before they reach a human reviewer.</p>
<h2 id="security-review-for-ai-generated-code-the-non-negotiable-step">Security Review for AI-Generated Code: The Non-Negotiable Step</h2>
<p>Security scanning of AI-generated code is not optional — it is the gate that separates acceptable production risk from negligence. The data is unambiguous: between 40 and 62% of AI-generated code contains security vulnerabilities, and the failure modes are predictable enough to be systematically caught before they ship. The two tools that belong in every agentic engineering CI pipeline in 2026 are Semgrep and Snyk Code. Semgrep performs static analysis against a continuously updated ruleset that covers OWASP Top 10 vulnerabilities, injection flaws, authentication bypasses, and insecure dependency usage — all categories that appear frequently in AI-generated code because the models learn from public repositories that include vulnerable patterns. Snyk Code adds a semantic layer that catches vulnerabilities Semgrep&rsquo;s pattern matching misses, including data flow issues that span multiple files and logic errors in access control implementations. The integration point for both tools is the CI pipeline: every pull request that includes AI-generated code runs Semgrep and Snyk Code before merge. Findings above a configured severity threshold block the merge. This is not a review burden — it is a force multiplier that catches the category of issues most likely to cause production incidents, before those incidents happen. Escape.tech scanned 5,600 vibe-coded applications and found 2,000 highly critical vulnerabilities and 400 exposed secrets. A CI-gated security scan catches the vast majority of those before they reach a URL.</p>
<h2 id="agent-orchestration-parallelism-and-incremental-commits">Agent Orchestration: Parallelism and Incremental Commits</h2>
<p>Effective agent orchestration requires two habits that run counter to vibe coding instincts: keeping individual agent tasks small and running independent tasks in parallel rather than sequentially. The single-mega-prompt antipattern — asking one agent to implement an entire feature in one shot — produces large, tangled diffs that are hard to review, hard to debug, and hard to roll back when something is wrong. The alternative is decomposing every feature into independent units that can be assigned to separate agent instances simultaneously. A REST endpoint implementation, for example, splits naturally into the route handler, the data validation layer, the business logic, the database query, and the test suite — five parallel tasks rather than one sequential monolith. Parallel agent execution compresses total wall-clock time without sacrificing reviewability. Each agent&rsquo;s output is a small, focused diff that a human reviewer can assess in minutes rather than hours. Incremental commits follow the same logic: every agent task that passes tests and security scan gets committed individually with a descriptive message. This creates a granular commit history that makes <code>git bisect</code> effective, rollbacks precise, and code review tractable. Teams that have standardized on this pattern report that their average pull request review time drops substantially because reviewers are never confronted with a 500-line diff that represents an hour of AI generation with no intermediate checkpoints. The orchestration overhead — decomposing tasks, launching parallel agents, reviewing incremental outputs — pays back immediately in reduced debugging time and faster review cycles.</p>
<h2 id="the-agentic-engineering-tool-stack-in-2026">The Agentic Engineering Tool Stack in 2026</h2>
<p>The tool stack for agentic engineering in 2026 is well-defined, with clear roles at each layer of the workflow. Starting with 80% of developers already using AI coding agents, the choice of primary agent interface shapes every other decision in the stack. Claude Code operates at the terminal and handles multi-step, multi-file tasks with explicit plan-then-execute separation — the <code>--plan</code> mode produces a reviewable action plan before the agent touches any files. Cursor 3 brings parallel agent execution inside an IDE interface with glass-panel agent windows that let you monitor multiple agent tasks simultaneously. Cline and Kilo Code extend VS Code with agent capabilities that integrate directly with existing editor workflows. At the security layer, Semgrep and Snyk Code handle static analysis and semantic vulnerability detection respectively, both with native GitHub Actions integration for CI enforcement. Langfuse provides observability: every agent call, token count, latency measurement, and output quality score is captured in a dashboard that makes agent performance visible and improvable over time. GitHub Actions ties the layers together — tests run on every commit, security scans run on every pull request, and deployment gates enforce that no AI-generated code reaches production without passing automated quality checks. The stack is not prescriptive at the primary agent layer — the principles of agentic engineering apply equally whether you prefer Claude Code, Cursor, or Copilot Workspace — but every production agentic workflow needs the observability, security, and CI layers regardless of which IDE-level agent you choose.</p>
<h2 id="building-a-repeatable-agentic-engineering-workflow">Building a Repeatable Agentic Engineering Workflow</h2>
<p>Repeatability is what separates a disciplined engineering practice from a series of lucky outcomes. The agentic engineering workflow becomes repeatable when every session follows the same sequence of steps and every deviation from that sequence is a conscious decision rather than an accident. The sequence has six phases: context initialization, task decomposition, test specification, parallel agent execution, automated gate review, and incremental commit. Context initialization means opening or updating the CLAUDE.md or AGENTS.md file before the first agent call of the session — the agent reads current project state before it writes a single line. Task decomposition means breaking the day&rsquo;s work into independent units that can be executed in parallel, each with a clear input, output, and success criterion. Test specification means writing the failing tests for each task before the agent generates implementation code. Parallel agent execution means launching multiple agent instances simultaneously for independent tasks, monitoring their outputs in real time, and intervening immediately when an agent drifts outside its assigned scope. Automated gate review means every agent output runs through the test suite and security scanner before it is considered complete — no exceptions for urgency or deadline pressure. Incremental commit means each passing task becomes its own commit, creating the granular history that makes future debugging and review tractable. Teams that have formalized this workflow report that the ramp-up cost for new developers drops significantly because the CLAUDE.md file and the commit history together provide enough context to understand any decision made in the codebase. The workflow is the documentation, because the workflow was designed to be legible from the start.</p>
<hr>
<h2 id="frequently-asked-questions">Frequently Asked Questions</h2>
<p><strong>Q1: What is the difference between agentic engineering and vibe coding?</strong></p>
<p>Vibe coding is prompt-driven, intuition-led development where the developer accepts AI-generated code without deep review. Agentic engineering is a structured workflow where the human engineer defines architecture, constraints, and validation criteria first, then orchestrates AI agents to execute specific tasks within those boundaries. The key difference is who holds engineering judgment: the AI in vibe coding, the human in agentic engineering. The distinction is invisible at the prototype stage and becomes critical in production systems.</p>
<p><strong>Q2: Why does vibe coding produce insecure code?</strong></p>
<p>AI models are trained on public repositories that include vulnerable code patterns. When a model generates code to satisfy a prompt, it can reproduce those patterns without flagging them as security issues. Between 40 and 62% of AI-generated code contains security vulnerabilities — injection flaws, authentication bypasses, insecure dependency usage, and exposed secrets are the most common categories. Vibe coding&rsquo;s lack of systematic review means these vulnerabilities ship without detection. Agentic engineering addresses this with CI-integrated static analysis tools like Semgrep and Snyk Code that run on every pull request before merge.</p>
<p><strong>Q3: What should a CLAUDE.md or AGENTS.md file contain?</strong></p>
<p>A production-grade context file should include: the project&rsquo;s tech stack and version constraints, the directory structure and module organization, naming conventions and code style rules, architectural decisions that are settled and should not be revisited, external APIs and services the project integrates with, security requirements and compliance constraints, and a summary of major decisions made in previous agent sessions. The file should be updated at the end of every significant agent session to capture new decisions. A well-maintained context file lets the agent start each session with full project awareness rather than making assumptions from scratch.</p>
<p><strong>Q4: How do I run AI agents in parallel without creating merge conflicts?</strong></p>
<p>Effective parallel agent execution requires decomposing tasks into genuinely independent units before launching agents. A task is independent if it touches a different set of files from every other currently running agent task. A REST endpoint, its tests, its database migration, and its documentation can all run in parallel because they modify separate files. When tasks do share files — for example, two features that both modify the same configuration file — they must run sequentially. The discipline of identifying dependencies before decomposing tasks is the core skill of agent orchestration. Tools like Cursor 3&rsquo;s parallel agent windows make the file-level activity of each agent visible, which makes conflict detection immediate rather than a surprise at merge time.</p>
<p><strong>Q5: What metrics should I track to measure the quality of my agentic engineering workflow?</strong></p>
<p>The five metrics that matter most are: (1) test pass rate on first agent attempt — how often does the agent&rsquo;s first implementation pass the tests you wrote before generation; (2) security scan findings per pull request — are vulnerabilities decreasing as your context file improves; (3) average pull request review time — smaller, incremental commits should reduce this significantly; (4) debugging time ratio — are you spending less time debugging AI-generated code than writing equivalent code manually; and (5) context file freshness — how many sessions ago was your CLAUDE.md last updated, and does it reflect the current state of the codebase. Langfuse tracks the agent-level metrics automatically; the others require instrumentation in your CI pipeline and a weekly review habit.</p>
]]></content:encoded></item><item><title>Vibe Coding vs Agentic Engineering: Which Workflow Is Right for You?</title><link>https://baeseokjae.github.io/posts/vibe-coding-vs-agentic-engineering-2026/</link><pubDate>Fri, 15 May 2026 18:10:54 +0000</pubDate><guid>https://baeseokjae.github.io/posts/vibe-coding-vs-agentic-engineering-2026/</guid><description>Vibe coding and agentic engineering both use AI, but the oversight model is completely different. Here&amp;#39;s how to choose the right workflow for your situation.</description><content:encoded><![CDATA[<p>Vibe coding lets AI write everything while you stay in &ldquo;the vibe,&rdquo; accepting code without deep review. Agentic engineering keeps a human engineer orchestrating AI agents — setting specs, reviewing outputs, and owning the final system. The right choice depends on what you&rsquo;re building, who will use it, and whether production failures are an option.</p>
<h2 id="what-is-vibe-coding-karpathys-original-definition">What Is Vibe Coding? Karpathy&rsquo;s Original Definition</h2>
<p>Vibe coding is a development approach coined by Andrej Karpathy in February 2025 where the developer fully delegates code generation to an AI model and accepts its output without detailed review — operating on intuition and iteration rather than engineering rigor. The term went mainstream fast: Collins English Dictionary named it Word of the Year for 2025, and by early 2026, 92% of US-based developers reported using some form of vibe coding in their workflows. The core mechanic is intentional surrender — you describe what you want in natural language, the AI generates code, you run it, and if it works well enough, you move on. There is no architecture phase, no design review, no systematic testing pass. Karpathy framed the style around accepting AI output even when you can&rsquo;t fully read or verify it, trusting the model&rsquo;s judgment over your own. This makes vibe coding extraordinarily fast for getting early prototypes to a visible, interactive state — 74% of developers using the approach report productivity increases and median task completion time drops 20–45% for greenfield features. The tradeoff is what happens next.</p>
<p>The practical workflow typically involves a conversational IDE like Lovable, Bolt, or Replit AI Agent where you describe features, the system generates full files, and you test in the browser. Cursor and Windsurf are also commonly used in vibe mode, where developers accept multi-file suggestions without reviewing individual diffs. The absence of deliberate review is a defining feature, not a bug — vibe coding optimizes for momentum and immediate feedback over correctness or maintainability. That&rsquo;s appropriate for some contexts and catastrophic for others.</p>
<h2 id="what-is-agentic-engineering-the-2026-evolution">What Is Agentic Engineering? The 2026 Evolution</h2>
<p>Agentic engineering is a structured development methodology, formalized by Karpathy on February 5, 2026, where a human engineer orchestrates one or more AI agents — defining specs, setting constraints, reviewing intermediate outputs, and validating final results — rather than passively accepting AI-generated code. The approach treats AI as a powerful but fallible junior contributor that needs architectural direction, clear task boundaries, and consistent review checkpoints. Unlike vibe coding&rsquo;s conversational flow, agentic engineering begins with a planning phase: what are the requirements, what are the constraints, what order should tasks execute in, and how will outputs be validated? An agentic pilot of 20+ debugging workflows at a production engineering team produced a 93% reduction in time-to-root-cause compared to historical baselines — that kind of outcome requires spec-driven orchestration, not vibes.</p>
<p>The agentic stack includes several layers that vibe coding skips entirely: a reasoning layer (how the agent plans), a memory layer (how context is preserved across multi-step tasks), a coordination layer (how parallel agents hand off work), a validation layer (automated and human review gates), and explicit human-in-the-loop checkpoints at defined stages. Tools designed for agentic engineering — Claude Code, Devin, GitHub Copilot Workspace — provide hooks for this oversight model. Gartner projects that 40% of enterprise applications will embed AI agents by end of 2026, and agentic engineering is the methodology that makes those deployments maintainable.</p>
<h2 id="core-differences-control-oversight-and-code-ownership">Core Differences: Control, Oversight, and Code Ownership</h2>
<table>
  <thead>
      <tr>
          <th>Dimension</th>
          <th>Vibe Coding</th>
          <th>Agentic Engineering</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>Code review</td>
          <td>Minimal to none</td>
          <td>Structured at checkpoints</td>
      </tr>
      <tr>
          <td>Planning phase</td>
          <td>None — conversational</td>
          <td>Explicit: spec, constraints, order</td>
      </tr>
      <tr>
          <td>Ownership model</td>
          <td>AI drives, human accepts</td>
          <td>Human orchestrates, AI executes</td>
      </tr>
      <tr>
          <td>Failure handling</td>
          <td>Retry with different prompt</td>
          <td>Root cause, fix spec, revalidate</td>
      </tr>
      <tr>
          <td>Testing</td>
          <td>Manual, ad hoc</td>
          <td>Automated gates + human sign-off</td>
      </tr>
      <tr>
          <td>Scale ceiling</td>
          <td>Prototype / solo project</td>
          <td>Production / team / enterprise</td>
      </tr>
      <tr>
          <td>Primary tools</td>
          <td>Lovable, Bolt, Replit</td>
          <td>Claude Code, Devin, Copilot Workspace</td>
      </tr>
  </tbody>
</table>
<p>The fundamental difference is not about which AI model you use — it&rsquo;s about who holds engineering judgment. In vibe coding, the AI holds it. In agentic engineering, the human does, and uses AI to accelerate execution of that judgment. This distinction matters more as system complexity grows. A single-page prototype lives or dies by whether it works right now. A production API that handles financial data, user authentication, or medical records lives or dies by whether it was designed to be correct and maintainable over time.</p>
<p>The trust gap is real and widening. 80% of developers now use AI coding agents in their workflows, yet trust in AI accuracy dropped from 40% to 29% year-over-year. Developers are using AI more while trusting it less — and agentic engineering is the workflow that makes that combination sustainable rather than dangerous.</p>
<h2 id="when-vibe-coding-works-and-when-it-breaks">When Vibe Coding Works (and When It Breaks)</h2>
<p>Vibe coding works best when the cost of failure is low, the scope is contained, and speed matters more than correctness. Specifically: personal tools that only you use, hackathon demos, internal dashboards with no sensitive data, one-off scripts for data transformation, and early-stage product prototypes where you need to validate a concept before investing engineering time. The 20–45% productivity boost is real and repeatable in these contexts. Founders and non-technical builders are especially effective with vibe coding for MVPs — non-developer adoption of AI coding tools surged 520% from 2024 to 2026 precisely because vibe coding removes the engineering barrier to getting something interactive in front of users.</p>
<p>Vibe coding breaks in four predictable patterns. First, security: 40–62% of AI-generated code contains security vulnerabilities, and AI-written code produces flaws at 2.74x the rate of human-written code. Georgia Tech&rsquo;s Vibe Security Radar tracked 35 CVEs in vibe-coded apps in March 2026 alone, up from 6 in January. Second, scale: the conversational approach generates code that solves the immediate prompt without concern for architecture, causing refactoring debt that compounds as the codebase grows. Third, debugging: 63% of developers have spent more time debugging AI-generated code than it would have taken to write themselves — vibe-coded systems have no specification to debug against, so failures are hard to isolate. Fourth, team collaboration: vibe-coded repositories often lack consistent patterns, documentation, or testable abstractions, making handoffs and code reviews nearly impossible.</p>
<h2 id="when-agentic-engineering-is-the-right-call">When Agentic Engineering Is the Right Call</h2>
<p>Agentic engineering is the right workflow when production correctness matters, when the system will be maintained by a team, when security or compliance is in scope, or when failure has real consequences for real users. The planning phase alone changes outcomes dramatically — by defining architecture before execution, you constrain the AI&rsquo;s solution space to approaches that are coherent, testable, and aligned with your non-functional requirements. Agentic engineering pilots showed a 65% reduction in development workflow execution time, with the biggest gains from compressing downstream testing — testing is cheaper when the spec is explicit and the agent&rsquo;s outputs are validated at each step rather than audited after the fact.</p>
<p>In practice, agentic engineering applies whenever you would normally write a technical spec. Authentication systems, payment processing, data pipelines, APIs consumed by multiple clients, systems with access control requirements, and any codebase that will be touched by more than one developer — these all benefit from the orchestration model. The overhead is real: you spend time writing specs, defining constraints, reviewing agent outputs, and setting up validation gates. That overhead pays back on the first production incident you prevent, the first security audit you pass without emergency patching, and the first developer who can pick up a section of the codebase without asking you what it does.</p>
<h2 id="real-world-evidence-security-risks-and-production-failures">Real-World Evidence: Security Risks and Production Failures</h2>
<p>The failure record for vibe coding in production is already extensive enough to be instructive. An Escape.tech scan of 5,600 vibe-coded applications found 2,000 highly critical vulnerabilities and 400 exposed secrets including API keys — the secrets exposure alone represents potential total compromise of every downstream service those apps touch. CVE-2025-48757 was filed against a Lovable-generated app with a broken authentication flow that allowed session hijacking. A Base44-generated application shipped with an auth bypass that gave any user admin access. These aren&rsquo;t outliers from bad prompts — they&rsquo;re the predictable result of using a workflow optimized for speed without the oversight layer that catches security issues.</p>
<p>The productivity paradox is equally important. METR&rsquo;s study of 16 experienced open-source developers across 246 real GitHub issues found that developers using AI tools were 19% slower on average than developers working without them. The slowdown came from debugging AI-generated code that looked correct but had subtle flaws, integrating AI-generated components that didn&rsquo;t fit the surrounding architecture, and rewriting sections that solved the prompt but not the underlying problem. These experienced developers were attempting to use AI tools like vibe coding — accepting outputs, iterating conversationally — in contexts that required the oversight model of agentic engineering. The workflow mismatch was the bug.</p>
<h2 id="tools-for-each-approach-from-lovable-to-claude-code">Tools for Each Approach: From Lovable to Claude Code</h2>
<table>
  <thead>
      <tr>
          <th>Tool</th>
          <th>Primary Paradigm</th>
          <th>Best For</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>Lovable</td>
          <td>Vibe coding</td>
          <td>Full-stack web prototypes, MVP validation</td>
      </tr>
      <tr>
          <td>Bolt</td>
          <td>Vibe coding</td>
          <td>Frontend-heavy prototypes, React apps</td>
      </tr>
      <tr>
          <td>Replit AI Agent</td>
          <td>Vibe coding</td>
          <td>Quick scripts, personal tools, education</td>
      </tr>
      <tr>
          <td>v0 (Vercel)</td>
          <td>Vibe coding</td>
          <td>UI component generation</td>
      </tr>
      <tr>
          <td>Cursor</td>
          <td>Hybrid</td>
          <td>Professional developers, any project size</td>
      </tr>
      <tr>
          <td>Windsurf</td>
          <td>Hybrid</td>
          <td>Multi-file features with some oversight</td>
      </tr>
      <tr>
          <td>Claude Code</td>
          <td>Agentic engineering</td>
          <td>Complex systems, production codebases</td>
      </tr>
      <tr>
          <td>Devin</td>
          <td>Agentic engineering</td>
          <td>Long-horizon tasks, automated PR workflow</td>
      </tr>
      <tr>
          <td>GitHub Copilot Workspace</td>
          <td>Agentic engineering</td>
          <td>Enterprise team workflows</td>
      </tr>
      <tr>
          <td>Cline</td>
          <td>Agentic engineering</td>
          <td>Agent pipelines, custom tool integration</td>
      </tr>
  </tbody>
</table>
<p>The hybrid tools — Cursor and Windsurf — are interesting because they can operate in either mode depending on how the developer uses them. Cursor&rsquo;s composer can be used as a vibe coding interface (describe it, accept everything) or as an agentic interface (define scope, review diffs, iterate on failures systematically). The tool doesn&rsquo;t enforce the workflow. The developer does. This is why tool choice alone doesn&rsquo;t determine outcomes — the mental model and review discipline matter more than the specific interface.</p>
<p>Claude Code occupies a distinct position because it&rsquo;s explicitly designed for the terminal-first, oversight-heavy workflow. You write specs, run targeted commands, review tool call outputs, and approve actions before they execute. The architecture assumes you want to know what the agent is doing and why. That assumption is the right one for production software, even when it slows the initial build.</p>
<h2 id="the-decision-framework-which-workflow-is-right-for-you">The Decision Framework: Which Workflow Is Right for You?</h2>
<p>Answer three questions to identify the right workflow:</p>
<p><strong>1. What happens when this breaks in production?</strong></p>
<ul>
<li>Nothing / minor inconvenience → vibe coding viable</li>
<li>User data exposure, revenue loss, security incident → agentic engineering required</li>
</ul>
<p><strong>2. Who else will touch this code?</strong></p>
<ul>
<li>Just you, one-off use → vibe coding viable</li>
<li>Team of 2+, or you in 6 months → agentic engineering required</li>
</ul>
<p><strong>3. What are the maintenance expectations?</strong></p>
<ul>
<li>Throwaway after demo / after feature validated → vibe coding viable</li>
<li>Ongoing product, iterative development → agentic engineering required</li>
</ul>
<p>If any of the three answers points to agentic engineering, use agentic engineering. The Karpathy framing is useful here: vibe coding raises the floor for what non-engineers can build; agentic engineering raises the ceiling for what professional engineers can ship. These aren&rsquo;t competing — they serve different people in different contexts.</p>
<p>A practical version for working engineers: use vibe coding in the first 20% of exploration (validating that something is technically feasible and what the UX should feel like) and switch to agentic engineering for the remaining 80% (building the version that ships). The prototype doesn&rsquo;t need to survive. The production system does.</p>
<h2 id="the-convergence-trend-are-they-merging-in-2026">The Convergence Trend: Are They Merging in 2026?</h2>
<p>The practical gap between vibe coding and agentic engineering is narrowing, but not in the direction that makes vibe coding safer for production — it&rsquo;s that vibe coders are gradually adopting more oversight practices as they experience production failures, and agentic tools are making structured workflows faster than they used to be. Simon Willison noted in May 2026 that tool capabilities in Cursor and Claude Code are pushing developers who started as vibe coders toward agentic patterns, because the tools surface failures in ways that demand systematic responses rather than &ldquo;try a different prompt.&rdquo;</p>
<p>The convergence is cultural as much as technical. The 520% surge in non-developer adoption brought a wave of vibe-native users who treat AI coding like a creative tool — generate, iterate, ship. Professional engineers who absorbed that culture are now seeing the production consequences and course-correcting. The correction typically looks like: adding code review to previously unreviewed AI outputs, writing specs before prompting rather than after, and using structured validation steps between agent tasks. Those are exactly the practices that define agentic engineering. The vocabulary may still say &ldquo;vibe,&rdquo; but the workflow is converging toward oversight.</p>
<p>The meaningful divergence that will persist: genuinely exploratory prototyping, personal tools, and education benefit from vibe coding&rsquo;s low friction. Production systems, team codebases, and security-sensitive applications will require agentic discipline regardless of how fast the tools get. Faster AI doesn&rsquo;t eliminate the need for human judgment at system boundaries — it just raises the cost of skipping it.</p>
<h2 id="bottom-line-for-developers-in-2026">Bottom Line for Developers in 2026</h2>
<p>The choice between vibe coding and agentic engineering isn&rsquo;t a matter of preference or skill level — it&rsquo;s a matter of what you&rsquo;re building and who bears the consequences if it breaks. Use vibe coding when failure is cheap and speed matters most: personal tools, early prototypes, hackathon demos, concept validation. Use agentic engineering when correctness is non-negotiable: production APIs, systems with user data, anything a team will maintain. Most professional engineers in 2026 will need fluency in both, with clear judgment about which context calls for which approach. The Karpathy framework is the clearest heuristic: vibe coding raises the floor for what non-engineers can build; agentic engineering raises the ceiling for what professional engineers can ship reliably.</p>
<p>GitHub reports 46% of all new code is now AI-generated, with Gartner projecting 60% by end of 2026. Enterprise adoption of AI coding tools grew 340% from 2024 to early 2026. The question is not whether AI writes your code — it does — but whether you&rsquo;re engineering the system around that AI with appropriate oversight, or just vibing with it and hoping production never reveals the gap between speed and correctness. The engineers who thrive are the ones who know when each approach applies and never confuse a prototype&rsquo;s workflow for a production system&rsquo;s requirements.</p>
<hr>
<h2 id="faq">FAQ</h2>
<p><strong>What is the difference between vibe coding and agentic engineering?</strong>
Vibe coding means delegating code generation to AI and accepting outputs without detailed review — optimized for speed and exploration. Agentic engineering means orchestrating AI agents with explicit specs, review checkpoints, and validation gates — optimized for correctness and maintainability. The fundamental difference is where engineering judgment lives: with the AI in vibe coding, with the human in agentic engineering.</p>
<p><strong>Who invented the terms vibe coding and agentic engineering?</strong>
Andrej Karpathy coined &ldquo;vibe coding&rdquo; in February 2025 to describe fully AI-delegated development without code review. He introduced &ldquo;agentic engineering&rdquo; on February 5, 2026 as a structured counterpart — the methodology of orchestrating AI agents while retaining engineering oversight. Both terms emerged from his observations about how developers were actually using AI coding tools in practice.</p>
<p><strong>Is vibe coding safe for production applications?</strong>
No — vibe coding is not appropriate for production applications that handle user data, authentication, payments, or any security-sensitive function. An Escape.tech scan of 5,600 vibe-coded apps found 2,000 highly critical vulnerabilities and 400 exposed secrets. AI-generated code produces security flaws at 2.74x the rate of human-written code, and without systematic review, those flaws ship.</p>
<p><strong>What tools are best for agentic engineering?</strong>
The leading tools for agentic engineering are Claude Code (terminal-first, oversight-focused), Devin (long-horizon autonomous tasks with PR workflow), GitHub Copilot Workspace (enterprise team integration), and Cline (custom agent pipelines). These tools are designed around the assumption that the developer wants visibility into what the agent is doing and control over what it executes.</p>
<p><strong>Can I use vibe coding and agentic engineering in the same project?</strong>
Yes — many experienced engineers use a hybrid approach: vibe coding in early exploration to validate concepts quickly, then switching to agentic engineering once they&rsquo;re building the version that ships. The prototype doesn&rsquo;t need the same rigor as the production system. The key discipline is knowing when to switch — usually when you&rsquo;ve decided the concept is worth building properly.</p>
]]></content:encoded></item></channel></rss>