The Future of Everything Is Lies, I Guess: Work

· ai coding business · Source ↗

TLDR

  • Aphyr argues that AI-at-work optimism ignores reliability, deskilling, labor-shock, and power-concentration risks that arrive when companies treat LLMs as employees.

Key Takeaways

  • The post is part of Aphyr’s longer “The Future of Everything is Lies” series and focuses on work, software, and labor markets.
  • Natural-language programming may become common, but Aphyr argues it does not behave like a compiler because ambiguous prompts do not preserve semantics.
  • LLM-assisted software can be useful, but high-stakes correctness still requires humans who can read, reason about, and verify the generated code.
  • The author compares “AI employees” to unreliable coworkers who agree, apologize, and produce work while still leaving hidden failure modes behind.
  • Automation risks include deskilling, monitoring fatigue, automation bias, and takeover hazards when humans lose the practice needed to intervene.
  • Aphyr is most worried about broad labor displacement and the way AI spending can move money from workers to cloud and model providers.

Why It Matters

  • The piece is useful because it pushes past “will AI code?” into the operational question of who stays accountable when generated work fails.
  • It frames AI adoption as a systems problem, not just a productivity story: reliability, incentives, supervision, and institutional power all matter.
  • For builders, the practical takeaway is to keep verification skills strong instead of letting prompt workflows become a substitute for understanding.

HN Discussion

  • Much of the thread focused on the page being blocked for UK readers because of the Online Safety Act, with commenters debating whether personal blogs with comments are actually covered.
  • Several readers pushed back on broad anti-executive framing, while others argued that AI-driven layoffs and power concentration are already visible.
  • The strongest technical discussion centered on whether LLM coding changes software engineering itself or simply adds another unreliable automation layer that still needs review.

Notable Comments

  • @monooso pointed to Ofcom’s checker and argued that article comments may fall under an exemption, while other commenters said the legal risk remains unsettled.
  • @greatpost challenged the post’s CEO framing and argued that company leadership quality varies.
  • @atomicnumber3 pushed back on the idea that starting a better company is a realistic answer for most workers.

Original | Discuss on HN