Study: Self-generated Agent Skills are useless

Published: (February 16, 2026 at 04:15 PM EST)
2 min read

Source: Hacker News

Authors

View PDF | HTML (experimental)

Abstract

Agent Skills are structured packages of procedural knowledge that augment LLM agents at inference time. Despite rapid adoption, there is no standard way to measure whether they actually help. We present SkillsBench, a benchmark of 86 tasks across 11 domains paired with curated Skills and deterministic verifiers. Each task is evaluated under three conditions: no Skills, curated Skills, and self‑generated Skills. We test 7 agent‑model configurations over 7,308 trajectories. Curated Skills raise average pass rate by 16.2 percentage points (pp), but effects vary widely by domain (+4.5 pp for Software Engineering to +51.9 pp for Healthcare) and 16 of 84 tasks show negative deltas. Self‑generated Skills provide no benefit on average, showing that models cannot reliably author the procedural knowledge they benefit from consuming. Focused Skills with 2–3 modules outperform comprehensive documentation, and smaller models with Skills can match larger models without them.

Subjects

  • Artificial Intelligence (cs.AI)

Citation

  • arXiv: 2602.12670 (cs.AI)
  • DOI: (arXiv‑issued DOI via DataCite, pending registration)

Submission history

From: Xiangyi Li
Version: v1 – Fri, 13 Feb 2026 07:06:06 UTC (1,366 KB)

0 views
Back to Blog

Related posts

Read more »

Chess engines do weird stuff

Training method Since AlphaZero, lc0‑style chess engines have been trained with reinforcement learning RL. Specifically, you have the engine search + model pla...