Code skimming. Command rewriting. Test, build, and git output compression. Token budget cascading. 17 languages. 14ms for 3,000 lines. Built in Rust.
Other tools filter terminal noise. Skim understands your code. It parses ASTs across 17 languages, strips implementation while preserving architecture, then optimizes every other type of context your agent consumes: test output, build errors, git diffs, and raw commands. 14ms for 3,000 lines. 48x faster on cache hits.
Context capacity is not the bottleneck. Attention is. Every token you send to an LLM dilutes its focus. Research consistently shows attention dilution in long contexts -- models lose track of critical details even within their window. More tokens means higher latency, degraded recall, and weaker reasoning. Past a threshold, adding context makes outputs worse. While other tools stop at filtering command output, Skim parses your actual code structure and optimizes the full spectrum of agent context: code, test output, build errors, git diffs, and commands. Deeper, broader, and smarter than anything else available.
Take a typical 80-file TypeScript project: 63,000 tokens. That contains maybe 5,000 tokens of actual signal. The rest is implementation noise the model doesn't need for architectural reasoning.
80% of the time, the model doesn't need implementation details. It doesn't care how you loop through users or validate emails. It needs to understand what your code does and how pieces connect.
That's where Skim comes in.
| Mode | Tokens | Reduction | Use Case |
|---|---|---|---|
| Full | 63,198 | 0% | Original source code |
| Structure | 25,119 | 60.3% | Understanding architecture |
| Signatures | 7,328 | 88.4% | API documentation |
| Types | 5,181 | 91.8% | Type system analysis |
For example:
// Before: Full implementation (100 tokens)
export function processUser(user: User): Result {
const validated = validateUser(user);
if (!validated) throw new Error("Invalid");
const normalized = normalizeData(user);
return await saveToDatabase(normalized);
}
// After: Structure only (12 tokens)
export function processUser(user: User): Result { /* ... */ }One command. 60-90% smaller. Your 63,000-token codebase? Now 5,000 tokens. Fits comfortably in a single prompt with room for your question.
That same 80-file project that wouldn't fit? Now you can ask: "Explain the entire authentication flow" or "How do these services interact?" β and the AI actually has enough context to answer.
- 17 languages including TypeScript, JavaScript, Python, Rust, Go, Java, C, C++, C#, Ruby, SQL, Kotlin, Swift, Markdown, JSON, YAML, TOML
- 6 transformation modes from full to minimal to pseudo to structure to signatures to types (15-95% reduction)
- 14.6ms for 3,000-line files. 48x faster on cache hits
- Token budget cascading that automatically selects the most aggressive mode fitting your budget
- Parallel processing with multi-file globs via rayon
- PreToolUse hook rewrites
cat,head,tail,cargo test,npm test,git diffinto skim equivalents - Two-layer rule system with declarative prefix-swap and custom argument handlers
- One command installs the hook for automatic, invisible context savings
- Parses and compresses output from cargo, go, vitest, jest, pytest
- Extracts failures, assertions, pass/fail counts while stripping noise
- Three-tier degradation from structured parse to regex fallback to passthrough
- Parses cargo, clippy, tsc build output
- Extracts errors, warnings, and summaries
- Parses ESLint, Ruff, mypy, golangci-lint output
- Extracts errors and warnings with severity grouping
- Three-tier degradation from structured parse to regex fallback to passthrough
- Parses npm, pnpm, pip, cargo audit/install/outdated output
- Extracts vulnerabilities, version conflicts, and dependency issues
skim git diff-- AST-aware: shows changed functions with full boundaries and+/-markers, strips diff noise--mode structureadds unchanged functions as signatures for context--mode fullshows entire files with change markers- Supports
--staged, commit ranges (HEAD~3,main..feature)
skim git show-- compressesgit showoutput in two modes:- Commit mode (default, e.g.
skim git show HEAD): strips commit header noise and renders the diff with the AST-aware pipeline - File-content mode (e.g.
skim git show HEAD:src/main.rs): applies skim's source transform to the file content - Three-tier degradation: AST-aware render β raw hunk render β guardrail passthrough
- Commit mode (default, e.g.
- Compresses
git statusandgit logwith flag-aware passthrough - All subcommands support
--jsonfor machine-readable output
skim discoverscans agent session history for optimization opportunitiesskim learndetects CLI error-retry patterns and generates correction rules- Output guardrail ensures compressed output is never larger than the original
npx rskim file.ts# Via Homebrew (macOS/Linux)
brew install dean0x/tap/skim
# Via npm
npm install -g rskim
# Via Cargo
cargo install rskimNote: Use
npxfor trying it out. For regular use, install globally to avoid npx overhead (~100-500ms per invocation).
git clone https://github.com/dean0x/skim.git
cd skim
cargo build --release
# Binary at target/release/skim# Try it with npx (no install)
npx rskim src/app.ts
# Or install globally for better performance
npm install -g rskim
# Extract structure from single file (auto-detects language)
skim src/app.ts
# Process entire directory recursively (auto-detects all languages)
skim src/
# Process current directory
skim .
# Process multiple files with glob patterns
skim 'src/**/*.ts'
# Process all TypeScript files with custom parallelism
skim '*.{js,ts}' --jobs 4
# Get only function signatures from multiple files
skim 'src/*.ts' --mode signatures --no-header
# Extract type definitions
skim src/types.ts --mode types
# Extract markdown headers (H1-H3 for structure, H1-H6 for signatures/types)
skim README.md --mode structure
# Pipe to other tools
skim src/app.ts | bat -l typescript
# Read from stdin (REQUIRES --language flag)
cat app.ts | skim - --language=typescript
# Override language detection for unusual file extensions
skim weird.inc --language=typescript
# Clear cache
skim --clear-cache
# Disable caching for pure transformation
skim file.ts --no-cache
# Show token reduction statistics
skim file.ts --show-stats# Basic usage (auto-detects language)
skim file.ts # Single file
skim src/ # Directory (recursive)
skim 'src/**/*.ts' # Glob pattern
# With options
skim file.ts --mode signatures # Different mode
skim src/ --jobs 8 # Parallel processing
skim - --language typescript # Stdin (requires --language)Common options:
-m, --mode- Transformation mode:structure(default),signatures,types,full,minimal,pseudo-l, --language- Override auto-detection (required for stdin only)-j, --jobs- Parallel processing threads (default: CPU cores)--no-cache- Disable caching--show-stats- Show token reduction stats--disable-analytics- Disable analytics recording
π Full Usage Guide β
Skim offers six modes with different levels of aggressiveness:
| Mode | Reduction | What's Kept | Use Case |
|---|---|---|---|
| Full | 0% | Everything (original source) | Testing/comparison |
| Minimal | 15-30% | All code, doc comments | Light cleanup |
| Pseudo | 30-50% | Logic flow, names, values | LLM context with logic |
| Structure | 70-80% | Signatures, types, classes, imports | Understanding architecture |
| Signatures | 85-92% | Only callable signatures | API documentation |
| Types | 90-95% | Only type definitions | Type system analysis |
skim file.ts --mode structure # Default
skim file.ts --mode pseudo # Pseudocode (strips types, visibility, decorators)
skim file.ts --mode signatures # More aggressive
skim file.ts --mode types # Most aggressive
skim file.ts --mode full # No transformationNote on JSON/YAML/TOML files: JSON, YAML, and TOML always use structure extraction regardless of mode. Since they are data (not code), there are no "signatures" or "types" to extractβonly structure. All modes produce identical output for these file types.
| Language | Status | Extensions | Notes |
|---|---|---|---|
| TypeScript | β | .ts, .tsx |
Excellent grammar |
| JavaScript | β | .js, .jsx |
Full ES2024 support |
| Python | β | .py, .pyi |
Complete coverage |
| Rust | β | .rs |
Up-to-date grammar |
| Go | β | .go |
Stable |
| Java | β | .java |
Good coverage |
| C | β | .c, .h |
Full C11 support |
| C++ | β | .cpp, .hpp, .cc, .hh, .cxx, .hxx |
C++20 support |
| Markdown | β | .md, .markdown |
Header extraction |
| JSON | β | .json |
Structure extraction (serde) |
| YAML | β | .yaml, .yml |
Multi-document support (serde) |
| TOML | β | .toml |
Structure extraction (toml) |
| C# | β | .cs |
Full grammar, structs/interfaces |
| Ruby | β | .rb |
Classes, modules, methods |
| SQL | β | .sql |
DDL/DML via tree-sitter-sequel |
| Kotlin | β | .kt, .kts |
Data classes, coroutines, sealed classes |
| Swift | β | .swift |
Protocols, generics, SwiftUI structs |
// Input
class UserService {
async findUser(id: string): Promise<User> {
const user = await db.users.findOne({ id });
if (!user) throw new NotFoundError();
return user;
}
}
// Output (structure mode)
class UserService {
async findUser(id: string): Promise<User> { /* ... */ }
}# Input
def process_data(items: List[Item]) -> Dict[str, Any]:
"""Process items and return statistics"""
results = {}
for item in items:
results[item.id] = calculate_metrics(item)
return results
# Output (structure mode)
def process_data(items: List[Item]) -> Dict[str, Any]: { /* ... */ }// Input
{
"user": {
"profile": {
"name": "Jane Smith",
"age": 28,
"tags": ["admin", "verified"]
},
"settings": {
"theme": "dark",
"notifications": true
}
},
"items": [
{"id": 1, "price": 100},
{"id": 2, "price": 200}
]
}
// Output (structure mode)
{
user: {
profile: {
name,
age,
tags
},
settings: {
theme,
notifications
}
},
items: {
id,
price
}
}# Input (Kubernetes manifests)
---
apiVersion: v1
kind: ConfigMap
metadata:
name: app-config
data:
database_url: postgres://localhost:5432
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: web-app
spec:
replicas: 3
# Output (structure mode)
apiVersion
kind
metadata:
name
data:
database_url
---
apiVersion
kind
metadata:
name
spec:
replicasπ More Examples (All Languages) β
Reduce codebase size by 60-90% to fit in LLM context windows:
skim src/ --no-header | llm "Analyze this codebase"
skim src/app.ts | llm "Review this architecture"Extract function signatures for documentation:
skim src/ --mode signatures > api-docs.txt
skim 'lib/**/*.py' --mode signatures > python-api.txtFocus on type definitions and interfaces:
skim src/ --mode types --no-header
skim 'src/**/*.ts' --mode typesQuick overview without implementation details:
skim large-file.py | less
skim src/auth/ | lessπ 10 Detailed Use Cases β
Skim includes a Skimmer plugin for Claude Code β a codebase orientation agent that maps project structure, finds task-relevant code, and generates integration plans.
Option A: Via the skim marketplace
/plugin marketplace add dean0x/skim
/plugin install skimmer
Option B: Direct from the standalone repo
/plugin marketplace add dean0x/skimmer
Note:
dean0x/skimis a custom marketplace. Unlike the official Claude Code plugin directory, custom marketplaces must be added explicitly before plugins become available.
# Orient for a specific task
/skim add JWT authentication
# General codebase orientation
/skim
The Skimmer agent uses rskim to extract code structure, then maps relevant files, signatures, and integration points for your task.
Caching is enabled by default for 40-50x faster repeated processing.
| Scenario | Time | Speedup |
|---|---|---|
| First run (no cache) | 244ms | 1.0x |
| Second run (cached) | 5ms | 48.8x faster! |
ls ~/.cache/skim/ # View cache
skim --clear-cache # Clear cache
skim file.ts --no-cache # Disable for one runHow it works:
- Cache key: SHA256(file path + mtime + mode)
- Automatic invalidation when files change
- Platform-specific cache directory
When to disable caching:
- One-time LLM transformations
- Stdin processing
- Disk-constrained environments
See exactly how much context you're saving with --show-stats:
skim file.ts --show-stats
# [skim] 1,000 tokens β 200 tokens (80.0% reduction)
skim 'src/**/*.ts' --show-stats
# [skim] 15,000 tokens β 3,000 tokens (80.0% reduction) across 50 file(s)Uses OpenAI's tiktoken (cl100k_base for GPT-3.5/GPT-4). Output to stderr for clean piping.
Skim automatically tracks token savings from every invocation in a local SQLite database (~/.cache/skim/analytics.db). View your savings with the stats subcommand:
skim stats # All-time dashboard (cost estimates always shown)
skim stats --since 7d # Last 7 days
skim stats --format json # Machine-readable output
skim stats --verbose # Include parse quality details
skim stats --clear # Reset analytics dataEnvironment variables:
| Variable | Description |
|---|---|
SKIM_DISABLE_ANALYTICS |
Set to 1, true, or yes to disable recording |
SKIM_INPUT_COST_PER_MTOK |
Override $/MTok for cost estimates (default: 3.0) |
SKIM_ANALYTICS_DB |
Override analytics database path |
Analytics recording is fire-and-forget (non-blocking) and does not affect command performance. Data is automatically pruned after 90 days.
Skim includes built-in DoS protections:
- Max recursion depth: 500 levels
- Max input size: 50MB per file
- Max AST nodes: 100,000 nodes
- Path traversal protection: Rejects malicious paths
- No code execution: Only parses, never runs code
π Security Details & Best Practices β π Vulnerability Disclosure β
Skim uses a clean, streaming architecture:
Language Detection β tree-sitter Parser β Transformation β Streaming Output
Design principles:
- Streaming-first: Output to stdout, no intermediate files
- Zero-copy: Uses
&strslices to minimize allocations - Error-tolerant: Handles incomplete/broken code gracefully
- Type-safe: Explicit error handling, no panics
π Architecture Deep Dive β
Target: <50ms for 1000-line files β Exceeded (14.6ms for 3000-line files)
| File Size | Lines | Time | Speed |
|---|---|---|---|
| Small | 300 | 1.3ms | 4.3Β΅s/line |
| Medium | 1500 | 6.4ms | 4.3Β΅s/line |
| Large | 3000 | 14.6ms | 4.9Β΅s/line |
Production TypeScript Codebase:
| Mode | Tokens | Reduction | LLM Context Multiplier |
|---|---|---|---|
| Full | 63,198 | 0% | 1.0x |
| Structure | 25,119 | 60.3% | 2.5x more code |
| Signatures | 7,328 | 88.4% | 8.6x more code |
| Types | 5,181 | 91.8% | 12.2x more code |
π Full Performance Benchmarks β
# Build and test
cargo build --release
cargo test --all-features
# Lint
cargo clippy -- -D warnings
cargo fmt -- --check
# Benchmark
cargo bench~30 minutes per language:
- Add tree-sitter grammar to
Cargo.toml - Update
Languageenum insrc/types.rs - Add file extension mapping
- Add test fixtures
- Run tests
Current: v2.4.1 β Stable
β Core β Code Reading (17 languages):
- TypeScript/JavaScript/Python/Rust/Go/Java/C/C++/C#/Ruby/SQL/Markdown/JSON/YAML/TOML
- 5 transformation modes: structure, signatures, types, minimal, full
- Token budget (
--tokens N), max lines (--max-lines N), last lines (--last-lines N) - Multi-file glob support, parallel processing, caching (40-50x speedup)
β Command Output Compression:
- Test runners: cargo test, pytest, vitest/jest, go test
- Build tools: cargo build, cargo clippy, tsc
- Git: status, diff, log
- File tools: find, ls, tree, grep, rg
- Log: JSON structured + plaintext dedup, debug filtering, stack trace collapsing
- Infra: gh, aws, curl, wget
- Three-tier degradation: Structured β Regex β Passthrough
β Agent Integration:
skim initβ hook installation for Claude Code, Cursor, Codex, Gemini, Copilot, OpenCodeskim rewriteβ command rewriting engine with--hookmode- MCP server mode for agent-native workflows
β Analytics & Intelligence:
skim statsβ persistent SQLite dashboard with cost estimationskim discoverβ missed optimization finder across agent sessionsskim learnβ CLI error pattern detection and correction rules
β Distribution:
- cargo (
cargo install rskim), npm (npx rskim), Homebrew (brew install dean0x/tap/skim) - 2,482 tests passing, 14.6ms performance (3x under target)
See CHANGELOG.md for version history.
Comprehensive guides for all aspects of Skim:
- π Usage Guide - Complete CLI reference and options
- π― Transformation Modes - Detailed mode comparison and examples
- π‘ Examples - Language-specific transformation examples
- π Use Cases - 10 practical scenarios with commands
- β‘ Caching - Caching internals and best practices
- π Security - DoS protections and security best practices
- ποΈ Architecture - System design and technical details
- β±οΈ Performance - Benchmarks and optimization guide
- π οΈ Development - Contributing and adding languages
| Tool | Role | What It Does |
|---|---|---|
| Skim | Context Optimization | Code-aware AST parsing across 17 languages, command rewriting, test/build/git output compression |
| DevFlow | Quality Orchestration | 18 parallel reviewers, working memory, self-learning, composable plugin system |
| Autobeat | Agent Orchestration | Autonomous orchestration. Eval loops, multi-agent pipelines, DAG dependencies, crash-proof persistence |
Skim optimizes every byte of context. DevFlow enforces production-grade quality. Autobeat scales execution across agents. No other stack covers all three.
Contributions welcome! Please:
- Check issues for existing work
- Open an issue to discuss major changes
- Follow existing code style (
cargo fmt,cargo clippy) - Add tests for new features
- Update documentation
π See Development Guide for detailed instructions
skim/
βββ crates/
β βββ rskim-core/ # Core library (language-agnostic)
β βββ rskim/ # CLI binary (I/O layer)
βββ tests/fixtures/ # Test files for each language
βββ benches/ # Performance benchmarks (planned)
MIT License - see LICENSE for details.
- tree-sitter - Fast, incremental parsing library
- clap - Command-line argument parsing
- ripgrep, bat, fd - Inspiration for Rust CLI design
- Website: https://dean0x.github.io/x/skim/
- Repository: https://github.com/dean0x/skim
- Issues: https://github.com/dean0x/skim/issues
- Crates.io: https://crates.io/crates/rskim
- npm: https://www.npmjs.com/package/rskim
Built with β€οΈ in Rust