LLMs consistently pick resumes they generate over ones by humans or other models

· ai business · Source ↗

TLDR

  • arXiv preprint finds LLMs self-prefer their own resume rewrites over human-written ones at 67-82% rates across commercial and open-source models.

Key Takeaways

  • Paper uses a controlled resume correspondence experiment: human resumes had executive summaries rewritten by an LLM, then a separate LLM instance scored them.
  • Self-preference bias ranged 67-82%; candidates using the same LLM as the evaluator were 23-60% more likely to be shortlisted.
  • Bias was largest in business-adjacent occupations like sales and accounting across 24 simulated hiring pipelines.
  • Interventions targeting LLMs’ self-recognition capabilities reduced the bias by more than 50%.
  • Paper calls for AI fairness frameworks to address AI-AI interaction biases, not just demographic disparities.

Hacker News Comment Review

  • Top commenter @hyperpape raised a methodological concern: the study may only show preference for LLM-rewritten summaries within a resume, not full LLM-generated resumes vs. human ones – a meaningful scope limitation.
  • Commenters broadly agreed the dynamic is self-reinforcing: training shapes output style, and the same training weights cause an evaluator instance to rate stylistically similar content higher, with no human signal in the loop.
  • A practical risk noted: organizations filtering with a given LLM will systematically recruit candidates who used that same LLM, potentially degrading hiring signal quality over time.

Notable Comments

  • @aykutseker: “The resume is no longer written for a human at all” – flags the structural feedback loop where both sides optimizing for the model removes human legibility entirely.
  • @ivansmf: argues the pattern mirrors auto-rater training where agent output is scored by the same agent, making self-preference an industry-wide training artifact, not just a hiring quirk.

Original | Discuss on HN