I saved 10M tokens (89%) on my Claude Code sessions with a CLI proxy #5848
muskiness
started this conversation in
1. Feature requests
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Feature Request: Built-in CLI output filtering/compression to reduce token usage (inspired by RTK)
Hey team! 👋
I recently came across an interesting project called [RTK (Rust Token Killer)](https://github.com/rtk-ai/rtk) — a CLI proxy that filters and compresses command output before it reaches the LLM context. The author shared their results on [Reddit](https://www.reddit.com/r/ClaudeAI/comments/1r2tt7q/i_saved_10m_tokens_89_on_my_claude_code_sessions/), and the numbers are impressive:
cargo test: 155 lines → 3 lines (98% reduction)git status: 119 chars → 28 chars (76% reduction)git log: compact summaries instead of full outputThe core idea is simple: most CLI output sent to the LLM is noise — passing tests, verbose logs, progress bars, redundant formatting. Stripping that out before it hits the context window saves a massive amount of tokens without losing any useful information.
Why this would be valuable as a built-in feature in your project:
Possible implementation scope:
git,npm,cargo,pip, test runners, etc.)I think this kind of optimization would be a huge quality-of-life improvement for users and a natural fit for your tool. Would love to hear your thoughts on whether this is something you'd consider exploring!
References:
Beta Was this translation helpful? Give feedback.
All reactions