Profling Claude Converstaions

Published: (April 17, 2026 at 04:11 AM EDT)
3 min read
Source: Dev.to

Source: Dev.to

Cover image for Profling Claude Converstaions

Overview

It is the ultimate flow‑killer. You sit down, open your IDE, get maybe three or four good turns into a complex refactor, and then—BAM.

“You have reached your message limit until 4:00 PM.”

It feels broken. You just started! How is the tank already empty? This happened to me so often—while burning through $50 worth of tokens a day—that I realized I was flying completely blind. I had no idea what was actually happening under the hood of my conversations.

I built Intern. It’s a tool to trace your Claude interactions and provide a heuristic‑based profile of exactly how you’re consuming resources.

Intern acts as a transparent proxy. You point your Claude code traffic through it, and it keeps a persistent history of every conversation. This allows you to go back and analyze exactly what happened in a session that nuked your limits.

The Heuristic Profile: A Reality Check

The core of the project is the profile command. It takes your raw conversation traces and applies a heuristic analysis to categorize the “work” being done.

When I profiled one of my “short” sessions that somehow cost me $18, this is what Intern surfaced:

=== Cost Report ===
MODEL                      MSGS   INPUT     OUTPUT    CACHE READ   TOTAL
claude-opus-4-6            233    $0.0400   $1.9850   $11.2288     $17.8246

=== Complexity Breakdown ===
COMPLEXITY  COUNT  %
mechanical  196    70.0%
reasoning   69     24.6%
trivial     15     5.4%

=== Tool Usage (322 total calls) ===
TOOL            COUNT   %      BAR
----            -----   -      ---
Bash            110     34.2%  █████████████████
Read            67      20.8%  ██████████
Edit            49      15.2%  ███████

What Intern actually shows you

  • Mechanical vs. Reasoning – 70 % of my messages were purely mechanical (file I/O, bash commands). I was using high‑tier reasoning limits for brick‑laying tasks.
  • The Cache Tax – $11.22 was spent just on Claude re‑reading context. Changing the system prompt or a large file mid‑way triggers an immediate financial and rate‑limit spike.
  • Trace Persistence – Intern saves everything to .jsonl files, letting you analyze patterns across weeks to identify “token hogs.”
  • Offload Candidates – The profiler flags “trivial” or “tool continuation” messages, suggesting they could be handled by a lighter model without sacrificing quality.

How to Profile Your Own Workflow

The setup is non‑intrusive and stays out of your way.

Install the Intern CLI tool

brew tap abhishekjha17/intern
brew install intern

For platforms other than macOS, refer to the installation guide in the repository.

Spin up the proxy

intern proxy
# Example output:
# 2026/04/17 13:29:08 intern proxy listening on :11411 → https://api.anthropic.com (traces → /Users/abhi17/.intern/traces/traces.jsonl)

Start Claude through Intern

export ANTHROPIC_BASE_URL=http://localhost:11411
claude

Intern will silently log every request and response while you work.

Run the heuristic report

When you hit a rate limit, run the analysis:

intern profile .intern/traces/traces.jsonl

The Choice

Tracing your conversation is about more than just costs; it’s about reclaiming control over your workflow’s path. Once you see which parts of your session are purely mechanical, you can consciously offload those branches to cheaper or local models. Instead of being forced down a single, expensive, rate‑limited road, you get to decide exactly which model is worth your context and when.

Project

abhishekjha17/intern

0 views
Back to Blog

Related posts

Read more »