Nick and Julian tackle the provocative question that’s dividing the tech world: should we be using LLMs to write production code? They explore the gap between code that “works” and code that’s maintainable, the surprising ways AI is reshaping the junior developer market, and why context windows matter less than you’d think. Plus: a detour into corporate personhood and whether software is fundamentally incompatible with capitalism.
Key Topics
Defining Vibe Coding
The spectrum from deliberate LLM use to pure vibe coding
Why most people conflate any LLM-assisted coding with vibe coding
Julian’s approach: granular prompting with manual review of every change
Nick’s front-end fluency vs. back-end experimentation
The Hidden Costs of LLM-Generated Code
Why the first shot often works but subsequent iterations fail
The “five different engineers” problem: inconsistent patterns and architecture
Real productivity math: 2 days design + 1 day vibe coding ≠ time saved when Julian spends 5+ days fixing it
The emotional cost of working in a codebase with no coherent structure
Context Windows: The Promise vs. Reality
Million-token context windows sound impressive, but hit practical limits fast
Why stuffing PDFs and user research into context doesn’t scale linearly
The human advantage: we’re constrained by time, not context window size
Retrieval Augmented Generation (RAG) as a workaround: loading only relevant context
The Junior Developer Crisis
Market dynamics: companies betting that 20-30% productivity gains from senior engineers replace junior hires
The pipeline problem: if we stop training juniors today, who will become the senior engineers tomorrow?
Short-sighted cost-cutting vs. long-term talent development
Current state: top CS graduates with debt and no job offers
Economic Viability Check
Julian’s estimate: 20-30% more productive, worth $20/month but not $10,000/month
The subsidization question: what happens when real costs emerge?
Factor of 9 rule: improvements need to be 9x better to overcome switching costs
LLMs are useful, but they are hitting diminishing returns on training improvements
Legal Liability in the AI Age
The Tea app disaster: vibe-coded dating app leaked 30,000 driver’s licenses
Liability doesn’t disappear with AI—you’re still accountable for what you ship
The airlock problem: compressing corporate liability layers into solo founders
Missing legal framework: should AI be able to earn money, execute contracts, and assume liability?
How LLMs Actually Work (And Why It Matters)
Next-token prediction and probability distributions
Context poisoning: why long, contradictory contexts flatten output quality
The “throw away context” strategy: starting fresh vs. course-correcting
Genesis prompts vs. iterative refinement
Best Use Cases Discovered
Test case generation: LLMs excel at creating comprehensive test coverage
Design placeholder content: realistic, varied fake data at scale
Error debugging: handling Node version conflicts and library incompatibility
Non-critical backend work that doesn’t need optimization
Key Takeaways
Vibe coding works for prototypes and personal projects, not production code that others must maintain. The time saved upfront gets consumed (and then some) when engineers have to refactor poorly architected code.
LLMs are 20-30% productivity boosters for experienced engineers, not replacements. The math only works at current subscription prices ($20/month), not at true compute costs.
The junior developer market collapse is short-sighted. Today’s cost savings create tomorrow’s talent shortage—LLMs aren’t good enough to replace the senior engineers who will retire in 5-10 years.
Context window size ≠ usable context. Million-token windows sound impressive, but practical limits hit much sooner due to file formats, content quality, and context poisoning.
Legal liability is the non-obvious barrier to AI-generated code at scale. We lack frameworks for who’s accountable when autonomous agents make costly mistakes.
Pull Quotes
Julian on vibe coding: “The LLM basically goes like, I just need to satisfy this and get it to this point, it doesn’t matter how I get there as long as it does the thing.”
Nick on the junior developer crisis: “If we’ve destroyed the pipeline of junior engineers that over time become the mid and senior architects, that’s a big problem.”
Julian on context poisoning: “The longer the context window gets and the more self-contradictory or nonsensical that the content within the context window gets, the flatter the distribution will get and the more likely you will be to get garbage output.”
On economic viability: “If I’m paying 20 bucks a month for that extra 20 to 30%, I definitely wouldn’t say it’s worth it [at $10,000/month]. I'd better be picking up some extra money from my job.”
Resources & Concepts Mentioned
Claude Code (command line tool for agentic coding)
Retrieval Augmented Generation (RAG)
Context windows: Claude Opus 4.1 (1M tokens ≈ 1,500 pages)
The Tea app data breach incident
The factor of 9 rule for overcoming switching costs
Teaser for Next Episode
“Is software fundamentally incompatible with capitalism?” — Julian drops a fire question about open source, value creation, and economic systems that demands its own episode.
Startup life is messy. Founders Julian Vergel de Dios and Nick Dazé share candid stories about building technology. From emerging trends and tech predictions to personal failures and philosophy, this podcast explores how entrepreneurship works.



