Star ๅކๅฒ่ถ‹ๅŠฟ
ๆ•ฐๆฎๆฅๆบ: GitHub API ยท ็”Ÿๆˆ่‡ช Stargazers.cn
README.md

caveman

why use many token when few do trick

Stars Last Commit License

Install โ€ข Benchmarks โ€ข Before/After โ€ข Intensity Levels โ€ข Compress โ€ข Why


A Claude Code skill/plugin and Codex plugin that makes agent talk like caveman โ€” cutting ~75% of output tokens while keeping full technical accuracy. Plus a companion tool that compresses your memory files to cut ~45% of input tokens every session.

Based on the viral observation that caveman-speak dramatically reduces LLM token usage without losing technical substance. So we made it a one-line install.

Before / After

๐Ÿ—ฃ๏ธ Normal Claude (69 tokens)

"The reason your React component is re-rendering is likely because you're creating a new object reference on each render cycle. When you pass an inline object as a prop, React's shallow comparison sees it as a different object every time, which triggers a re-render. I'd recommend using useMemo to memoize the object."

๐Ÿชจ Caveman Claude (19 tokens)

"New object ref each render. Inline object prop = new ref = re-render. Wrap in useMemo."

๐Ÿ—ฃ๏ธ Normal Claude

"Sure! I'd be happy to help you with that. The issue you're experiencing is most likely caused by your authentication middleware not properly validating the token expiry. Let me take a look and suggest a fix."

๐Ÿชจ Caveman Claude

"Bug in auth middleware. Token expiry check use < not <=. Fix:"

Same fix. 75% less word. Brain still big.

Sometimes too much caveman. Sometimes not enough:

๐Ÿชถ Lite

"Your component re-renders because you create a new object reference each render. Inline object props fail shallow comparison every time. Wrap it in useMemo."

๐Ÿชจ Full

"New object ref each render. Inline object prop = new ref = re-render. Wrap in useMemo."

๐Ÿ”ฅ Ultra

"Inline obj prop โ†’ new ref โ†’ re-render. useMemo."

Same answer. You pick how many word.

Benchmarks

Real token counts from the Claude API (reproduce it yourself):

TaskNormal (tokens)Caveman (tokens)Saved
Explain React re-render bug118015987%
Fix auth middleware token expiry70412183%
Set up PostgreSQL connection pool234738084%
Explain git rebase vs merge70229258%
Refactor callback to async/await38730122%
Architecture: microservices vs monolith44631030%
Review PR for security issues67839841%
Docker multi-stage build104229072%
Debug PostgreSQL race condition120023281%
Implement React error boundary345445687%
Average121429465%

Range: 22%โ€“87% savings across prompts.

[!IMPORTANT] Caveman only affects output tokens โ€” thinking/reasoning tokens are untouched. Caveman no make brain smaller. Caveman make mouth smaller. Biggest win is readability and speed, cost savings are a bonus.

Science back caveman up

A March 2026 paper "Brevity Constraints Reverse Performance Hierarchies in Language Models" found that constraining large models to brief responses improved accuracy by 26 percentage points on certain benchmarks and completely reversed performance hierarchies. Verbose not always better. Sometimes less word = more correct.

Install

npx skills add JuliusBrussee/caveman

npx skills supports 40+ agents โ€” Claude Code, GitHub Copilot, Cursor, Windsurf, Cline, and more. To install for a specific agent:

npx skills add JuliusBrussee/caveman -a cursor npx skills add JuliusBrussee/caveman -a github-copilot npx skills add JuliusBrussee/caveman -a cline npx skills add JuliusBrussee/caveman -a windsurf npx skills add JuliusBrussee/caveman -a codex

Or with Claude Code plugin system:

claude plugin marketplace add JuliusBrussee/caveman claude plugin install caveman@caveman

Usage

Trigger with:

  • /caveman or Codex $caveman
  • "talk like caveman"
  • "caveman mode"
  • "less tokens please"

Stop with: "stop caveman" or "normal mode"

Intensity Levels

Sometimes full caveman too much. Sometimes not enough. Now you pick:

LevelTriggerWhat it do
Lite/caveman lite or $caveman liteDrop filler, keep grammar. Professional but no fluff
Full/caveman full or $caveman fullDefault caveman. Drop articles, fragments, full grunt
Ultra/caveman ultra or $caveman ultraMaximum compression. Telegraphic. Abbreviate everything

Level stick until you change it or session end.

What Caveman Do

ThingCaveman Do?
English explanation๐Ÿชจ Caveman smash filler words
Code blocksโœ๏ธ Write normal (caveman not stupid)
Technical terms๐Ÿง  Keep exact (polymorphism stay polymorphism)
Error messages๐Ÿ“‹ Quote exact
Git commits & PRsโœ๏ธ Write normal
Articles (a, an, the)๐Ÿ’€ Gone
Pleasantries๐Ÿ’€ "Sure I'd be happy to" is dead
Hedging๐Ÿ’€ "It might be worth considering" extinct

Why

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚  TOKENS SAVED          โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ 75% โ”‚
โ”‚  TECHNICAL ACCURACY    โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ 100%โ”‚
โ”‚  SPEED INCREASE        โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ ~3x โ”‚
โ”‚  VIBES                 โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ OOG โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜
  • Faster response โ€” less token to generate = speed go brrr
  • Easier to read โ€” no wall of text, just the answer
  • Same accuracy โ€” all technical info kept, only fluff removed (science say so)
  • Save money โ€” ~71% less output token = less cost
  • Fun โ€” every code review become comedy

How It Work

Caveman not dumb. Caveman efficient.

Normal LLM waste token on:

  • "I'd be happy to help you with that" (8 wasted tokens)
  • "The reason this is happening is because" (7 wasted tokens)
  • "I would recommend that you consider" (7 wasted tokens)
  • "Sure, let me take a look at that for you" (10 wasted tokens)

Caveman say what need saying. Then stop.

Caveman Compress

Caveman makes Claude speak with fewer tokens. Caveman Compress makes Claude read fewer tokens.

Your CLAUDE.md loads on every session start. A 1000-token project memory file costs you tokens every single time you open a project. Caveman Compress rewrites those files into caveman-speak so Claude reads less โ€” without you losing the human-readable original.

/caveman-compress CLAUDE.md
CLAUDE.md          โ† compressed (Claude reads this every session โ€” fewer tokens)
CLAUDE.original.md โ† human-readable backup (you read and edit this)

How it works

A Python pipeline that shells out to claude --print for the actual compression, then validates the result locally โ€” no tokens wasted on checking.

detect file type (local)  โ†’  compress with Claude (1 call)  โ†’  validate (local)
                                                                    โ†“
                                                              if errors: targeted fix (1 call, cherry-pick only)
                                                                    โ†“
                                                              retry up to 2ร—, restore original on failure

What's preserved exactly

Code blocks, inline code, URLs, file paths, commands, headings, table structure, dates, version numbers โ€” anything technical passes through untouched. Only natural language prose gets compressed.

Compress benchmarks

FileOriginalCompressedSaved
claude-md-preferences.md70628559.6%
project-notes.md114553553.3%
claude-md-project.md112268738.8%
todo-list.md62738838.1%
mixed-with-code.md88857435.4%
Average89849445%

Full-circle token savings

ToolWhat it cutsSavings
cavemanOutput tokens (Claude's responses)~65%
caveman-compressInput tokens (memory files loaded per session)~45%
Both togetherThe whole conversationOutput + input both shrunk

See the full caveman-compress README for install, usage, and validation details.

Star This Repo

If caveman save you mass token, mass money โ€” leave mass star. โญ

Star History Chart

Also by Julius Brussee

  • Blueprint โ€” specification-driven development for Claude Code. Natural language โ†’ blueprints โ†’ parallel builds โ†’ working software.
  • Revu โ€” local-first macOS study app with FSRS spaced repetition, decks, exams, and study guides. revu.cards

License

MIT โ€” free like mass mammoth on open plain.

ๅ…ณไบŽ About

๐Ÿชจ why use many token when few token do trick โ€” Claude Code skill that cuts 65% of tokens by talking like caveman
aianthropiccavemanclaudeclaude-codellmmemeprompt-engineeringskilltokens

่ฏญ่จ€ Languages

Python100.0%

ๆไบคๆดป่ทƒๅบฆ Commit Activity

ไปฃ็ ๆไบค็ƒญๅŠ›ๅ›พ
่ฟ‡ๅŽป 52 ๅ‘จ็š„ๅผ€ๅ‘ๆดป่ทƒๅบฆ
32
Total Commits
ๅณฐๅ€ผ: 32ๆฌก/ๅ‘จ
Less
More

ๆ ธๅฟƒ่ดก็Œฎ่€… Contributors