A high-performance, open-source AI coding agent written in Rust.
ClaudeCode is a terminal-native CLI agent designed to bring advanced LLM capabilities directly into your development workflow. Built for speed, safety, and efficiency, it provides an interactive agent shell, workspace-aware tools, and persistent session management. It is an independent open-source implementation inspired by Claude Code, not the official Anthropic product.
- Rust-powered: Built with Rust for memory safety, minimal binary size, and high execution speed.
- Agentic CLI: Interactive shell and one-shot prompt support for seamless terminal workflows.
- Model flexible: Supports Anthropic-compatible and OpenAI-compatible providers, plus xAI/Grok aliases.
- Workspace aware: Context-aware tools designed to understand your local codebase.
- Session persistence: Resumeable sessions via JSON state management.
- Extensible: Plugin-ready architecture for custom tools and skills.
- Install Rust stable and Cargo.
- Set up your preferred API credentials.
From the repository root:
cd rust
cargo build --release -p claw-cli
# Install locally to your PATH for global access
cargo install --path crates/claw-cli --lockedStart the interactive shell:
clawRun a single prompt:
claw prompt "summarize this workspace"Resume a previous session:
claw --resume session.json /statusRun claw --help for the full command list, including agents, skills, and system-prompt flows.
Configure your environment variables based on your preferred provider:
export ANTHROPIC_API_KEY="..."
export ANTHROPIC_BASE_URL="https://api.anthropic.com"export OPENAI_API_KEY="..."
export OPENAI_BASE_URL="https://api.openai.com/v1"export XAI_API_KEY="..."
export XAI_BASE_URL="https://api.x.ai"You can also authenticate via the CLI:
claw loginWhat is this project? This is an independent, open-source implementation of a terminal-based coding agent, architecturally inspired by Claude Code.
Why Rust? Rust provides the performance, concurrency, and memory safety required for a tool that interacts deeply with local file systems and high-latency LLM APIs.
Can I use local models? Yes, if your local inference server exposes an OpenAI-compatible API and you point the relevant base URL and API key at it.
Is this the official Anthropic Claude Code? No, this is a community-driven open-source project.
We welcome contributions. Please refer to CLAW.md for workspace-specific workflow guidance.
cd rust
cargo fmt
cargo clippy --workspace --all-targets -- -D warnings
cargo test --workspacerust/: Core CLI and runtime implementation.src/: Python support code and utilities.tests/: Verification suites for agentic behaviors.CLAW.md: Internal workflow documentation.
- This project is an open-source implementation.
- It is not affiliated with or endorsed by Anthropic.