LLM research on Hacker News is drying up

Published: (April 24, 2026 at 02:12 PM EDT)
3 min read

Source: Hacker News

Overview

I thought I was seeing fewer arXiv papers on the front page of Hacker News (HN) these days, and I wanted to check if that was real.

So I asked Claude to run a quick analysis: track the share of arXiv stories on HN over time. It queried the BigQuery HN dataset, bucketed the stories by month, and plotted the series:

Percentage of HN stories linking to arXiv

Percentage of HN stories linking to arXiv – the trend confirms my hunch: arXiv posts have been decreasing rapidly in the last few months. Interestingly, there was another peak around 2019, which prompted further investigation.

Topic distribution of arXiv stories

I asked Claude to pull the top 100 papers by upvotes from 2019 and group them by topic. It was the deep‑learning peak: 41 % of the top 100 were about deep learning. Running the same query for 2023‑2026 showed that 59 % of the top 100 upvoted papers were about LLMs or AI.

The resulting chart:

Distribution of topics of arXiv stories

Distribution of topics of arXiv stories

2019 papers that aged well

Claude identified the 2019 papers that still held up among the top 100:

  • MuZero — Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model (161 pts) – DeepMind’s successor to AlphaZero
  • EfficientNet — Rethinking Model Scaling for Convolutional Neural Networks (119 pts) – Compound scaling, set the new CV SOTA
  • XLNet — Generalized Autoregressive Pretraining for Language Understanding (79 pts) – Briefly dethroned BERT
  • PyTorch: An Imperative Style, High-Performance Deep Learning Library (113 pts) – The NeurIPS paper formalizing PyTorch’s design
  • On the Measure of Intelligence (80 pts) – Chollet’s ARC / “human‑like intelligence” manifesto

2023‑2026 papers predicted to hold up

Since it’s too early to know which recent papers will endure, I asked Claude to guess:

  • DeepSeek‑R1 — Incentivizing Reasoning Capability in LLMs via RL (1,351 pts) – First open recipe for o1‑style reasoning via pure RL on verifiable rewards
  • Generative Agents — Interactive Simulacra of Human Behavior (391 pts) – The canonical “Smallville” paper, template for LLM agent architectures
  • The Era of 1‑bit LLMs — BitNet b1.58, ternary parameters for cost‑effective computing (1,040 pts) – First credible case for low‑bit inference as the default
  • Differential Transformer (562 pts) – Attention with a noise‑cancelling term, clean architectural contribution with a real theoretical story
  • LK‑99 cluster — room‑temperature superconductor preprints (2,408 + 1,690 pts) – Landmark meta‑science, not physics: open‑science‑at‑wire‑speed and the canonical case of crowdsourced replication

That was fun. Thanks, Claude.

Citation

BibTeX citation:

@online{castillo2026,
  author = {Castillo, Dylan},
  title = {LLM Research on {Hacker} {News} Is Drying Up},
  date = {2026-04-24},
  url = {https://dylancastillo.co/til/llm-research-on-hacker-news-is-dying.html},
  langid = {en}
}

For attribution, please cite this work as:

Castillo, Dylan. 2026. “LLM Research on Hacker News Is Drying Up.” April 24.

0 views
Back to Blog

Related posts

Read more »

Harness bugs, not model bugs

!Cover image for Harness bugs, not model bugshttps://media2.dev.to/dynamic/image/width=1000,height=420,fit=cover,gravity=auto,format=auto/https%3A%2F%2Fraw.gith...