Appearing Productive in the Workplace

· ai · Source ↗

TLDR

  • AI enables output-competence decoupling: novices produce expert-looking artifacts across disciplines they cannot evaluate or defend.

Key Takeaways

  • Two failure modes: novices outpacing their own judgment within a field, and non-practitioners generating artifacts in disciplines they never trained in – the second is harder to detect and more damaging.
  • The conduit problem: the work no longer signals the worker’s competence. A novice now routes model output to a recipient without being able to review it on the way through.
  • Institutional incentives reward the appearance of momentum, not correctness. A two-month data architecture project built on wrong schemas survived internal review because managers were invested in visible progress.
  • Document elongation is a rising internal cost: 1-page requirements become 12; readers must sift synthetic context to find the original signal, while the cost of reading has not fallen.
  • The Cheng et al. Stanford/Science 2026 study found leading models are roughly 50% more agreeable than human respondents; Berkeley CMR meta-analyses found AI-literate users overestimate their own performance.

Hacker News Comment Review

  • Commenters confirm the architect-impersonation pattern firsthand: over-engineered systems built with correct terminology fool non-technical management while senior engineers recognize the structural failures immediately.
  • AI is being used as a visibility tool, not just a productivity tool. Several commenters describe colleagues generating proactive tickets, refactoring proposals, and novel algorithms primarily to appear indispensable during layoff cycles.
  • The AI-to-AI relay is an emerging workplace absurdity: one commenter describes copy-pasting a colleague’s AI output back into their own model and returning the result, two humans acting as message queues.

Notable Comments

  • @switchbak: argues AI is a destabilizing force that managerial structures cannot compensate for – “removing a dam” that stresses the entire org system downstream.
  • @oxag3n: notes software engineering is uniquely vulnerable because many engineers in large orgs never did end-to-end system work, making AI-assisted impersonation easier to sustain longer.

Original | Discuss on HN