ruflo
Multi-agent orchestration MCP server for Claude Code — deploys swarms, coordinates workflows, learns from outcomes
Ruflo is an enterprise-grade multi-agent orchestration MCP server that transforms Claude Code into a powerful platform for deploying and coordinating specialized AI agent swarms. It features a self-learning architecture that optimizes task routing, manages agent consensus, and integrates with various LLM providers. Native integration via Model Context Protocol allows direct command execution within Claude Code sessions, enhancing development workflows.
- Deploy 100+ specialized AI agents for coding, review, and DevOps
- Coordinate unlimited agents in swarms with hierarchical or mesh patterns
- Self-learning architecture remembers patterns, optimizes task routing
- Integrates with Claude, GPT, Gemini, Cohere, and local LLM providers
- Native integration with Claude Code via Model Context Protocol (MCP)
README
View on GitHub ↗🌊 RuFlo v3.5: Enterprise AI Orchestration Platform

Multi-agent AI orchestration for Claude Code
Deploy 16 specialized agent roles + custom types in coordinated swarms with self-learning capabilities, fault-tolerant consensus, and enterprise-grade security.
Why Ruflo? Claude Flow is now Ruflo — named by Ruv, who loves Rust, flow states, and building things that feel inevitable. The "Ru" is the Ruv. The "flo" is the flow. Underneath, WASM kernels written in Rust power the policy engine, embeddings, and proof system. 6,000+ commits later, this is v3.5.
Getting into the Flow
Ruflo is a comprehensive AI agent orchestration framework that transforms Claude Code into a powerful multi-agent development platform. It enables teams to deploy, coordinate, and optimize specialized AI agents working together on complex software engineering tasks.
Self-Learning/Self-Optimizing Agent Architecture
User → Ruflo (CLI/MCP) → Router → Swarm → Agents → Memory → LLM Providers
↑ ↓
└──── Learning Loop ←──────┘
📐 Expanded Architecture — Full system diagram with RuVector intelligence
flowchart TB
subgraph USER["👤 User Layer"]
U[User]
end
subgraph ENTRY["🚪 Entry Layer"]
CLI[CLI / MCP Server]
AID[AIDefence Security]
end
subgraph ROUTING["🧭 Routing Layer"]
QL[Q-Learning Router]
MOE[MoE - 8 Experts]
SK[Skills - 130+]
HK[Hooks - 27]
end
subgraph SWARM["🐝 Swarm Coordination"]
TOPO[Topologies<br/>mesh/hier/ring/star]
CONS[Consensus<br/>Raft/BFT/Gossip]
CLM[Claims<br/>Human-Agent Coord]
end
subgraph AGENTS["🤖 100+ Agents"]
AG1[coder]
AG2[tester]
AG3[reviewer]
AG4[architect]
AG5[security]
AG6[...]
end
subgraph RESOURCES["📦 Resources"]
MEM[(Memory<br/>AgentDB)]
PROV[Providers<br/>Claude/GPT/Gemini/Ollama]
WORK[Workers - 12<br/>ultralearn/audit/optimize]
end
subgraph RUVECTOR["🧠 RuVector Intelligence Layer"]
direction TB
subgraph ROW1[" "]
SONA[SONA<br/>Self-Optimize<br/><0.05ms]
EWC[EWC++<br/>No Forgetting]
FLASH[Flash Attention<br/>2.49-7.47x]
end
subgraph ROW2[" "]
HNSW[HNSW<br/>HNSW-indexed]
RB[ReasoningBank<br/>Pattern Store]
HYP[Hyperbolic<br/>Poincaré]
end
subgraph ROW3[" "]
LORA[LoRA/Micro<br/>low-rank adaptation]
QUANT[Int8 Quant<br/>3.92x memory]
RL[9 RL Algos<br/>Q/SARSA/PPO/DQN]
end
end
subgraph LEARNING["🔄 Learning Loop"]
L1[RETRIEVE] --> L2[JUDGE] --> L3[DISTILL] --> L4[CONSOLIDATE] --> L5[ROUTE]
end
U --> CLI
CLI --> AID
AID --> QL & MOE & SK & HK
QL & MOE & SK & HK --> TOPO & CONS & CLM
TOPO & CONS & CLM --> AG1 & AG2 & AG3 & AG4 & AG5 & AG6
AG1 & AG2 & AG3 & AG4 & AG5 & AG6 --> MEM & PROV & WORK
MEM --> SONA & EWC & FLASH
SONA & EWC & FLASH --> HNSW & RB & HYP
HNSW & RB & HYP --> LORA & QUANT & RL
LORA & QUANT & RL --> L1
L5 -.->|loops back| QL
style RUVECTOR fill:#1a1a2e,stroke:#e94560,stroke-width:2px
style LEARNING fill:#0f3460,stroke:#e94560,stroke-width:2px
style USER fill:#16213e,stroke:#0f3460
style ENTRY fill:#1a1a2e,stroke:#0f3460
style ROUTING fill:#1a1a2e,stroke:#0f3460
style SWARM fill:#1a1a2e,stroke:#0f3460
style AGENTS fill:#1a1a2e,stroke:#0f3460
style RESOURCES fill:#1a1a2e,stroke:#0f3460
RuVector Components (included with Ruflo):
| Component | Purpose | Performance |
|---|---|---|
| SONA | Self-Optimizing Pattern Learning - learns optimal routing | Fast adaptation |
| EWC++ | Elastic Weight Consolidation - prevents catastrophic forgetting | Preserves learned patterns |
| Flash Attention | Optimized attention computation | 2-7x speedup (benchmarked) |
| HNSW | Hierarchical Navigable Small World vector search | Sub-millisecond retrieval |
| ReasoningBank | Pattern storage with trajectory learning | RETRIEVE→JUDGE→DISTILL |
| Hyperbolic | Poincare ball embeddings for hierarchical data | Better code relationships |
| LoRA/MicroLoRA | Low-Rank Adaptation for efficient fine-tuning | Lightweight adaptation |
| Int8 Quantization | Memory-efficient weight storage | ~4x memory reduction |
| SemanticRouter | Semantic task routing with cosine similarity | Fast intent routing |
| 9 RL Algorithms | Q-Learning, SARSA, A2C, PPO, DQN, Decision Transformer, etc. | Task-specific learning |
# Use RuVector via Ruflo
npx ruflo@latest hooks intelligence --status
Get Started Fast
# One-line install (recommended)
curl -fsSL https://cdn.jsdelivr.net/gh/ruvnet/ruflo@main/scripts/install.sh | bash
# Or full setup with MCP + diagnostics
curl -fsSL https://cdn.jsdelivr.net/gh/ruvnet/ruflo@main/scripts/install.sh | bash -s -- --full
# Or via npx
npx ruflo@latest init --wizard
New to Ruflo? You don't need to learn 310+ MCP tools or 26 CLI commands. After running
init, just use Claude Code normally — the hooks system automatically routes tasks to the right agents, learns from successful patterns, and coordinates multi-agent work in the background. The advanced tools exist for fine-grained control when you need it.
Key Capabilities
🤖 100+ Specialized Agents - Ready-to-use AI agents for coding, code review, testing, security audits, documentation, and DevOps. Each agent is optimized for its specific role.
🐝 Coordinated Agent Teams - Run unlimited agents simultaneously in organized swarms. Agents spawn sub-workers, communicate, share context, and divide work automatically using hierarchical (queen/workers) or mesh (peer-to-peer) patterns.
🧠 Learns From Your Workflow - The system remembers what works. Successful patterns are stored and reused, routing similar tasks to the best-performing agents. Gets smarter over time.
🔌 Works With Any LLM - Switch between Claude, GPT, Gemini, Cohere, or local models like Llama. Automatic failover if one provider is unavailable. Smart routing picks the cheapest option that meets quality requirements.
⚡ Plugs Into Claude Code - Native integration via MCP (Model Context Protocol). Use ruflo commands directly in your Claude Code sessions with full tool access.
🔒 Production-Ready Security - Built-in protection against prompt injection, input validation, path traversal prevention, command injection blocking, and safe credential handling.
🧩 Extensible Plugin System - Add custom capabilities with the plugin SDK. Create workers, hooks, providers, and security modules. Share plugins via the decentralized IPFS marketplace.
A multi-purpose Agent Tool Kit
🔄 Core Flow — How requests move through the system
Every request flows through four layers: from your CLI or Claude Code interface, through intelligent routing, to specialized agents, and finally to LLM providers for reasoning.
| Layer | Components | What It Does |
|---|---|---|
| User | Claude Code, CLI | Your interface to control and run commands |
| Orchestration | MCP Server, Router, Hooks | Routes requests to the right agents |
| Agents | 100+ types | Specialized workers (coder, tester, reviewer...) |
| Providers | Anthropic, OpenAI, Google, Ollama | AI models that power reasoning |
🐝 Swarm Coordination — How agents work together
Agents organize into swarms led by queens that coordinate work, prevent drift, and reach consensus on decisions—even when some agents fail.
| Layer | Components | What It Does |
|---|---|---|
| Coordination | Queen, Swarm, Consensus | Manages agent teams (Raft, Byzantine, Gossip) |
| Drift Control | Hierarchical topology, Checkpoints | Prevents agents from going off-task |
| Hive Mind | Queen-led hierarchy, Collective memory | Strategic/tactical/adaptive queens coordinate workers |
| Consensus | Byzantine, Weighted, Majority | Fault-tolerant decisions (2/3 majority for BFT) |
Hive Mind Capabilities:
- 🐝 Queen Types: Strategic (planning), Tactical (execution), Adaptive (optimization)
- 👷 8 Worker Types: Researcher, Coder, Analyst, Tester, Architect, Reviewer, Optimizer, Documenter
- 🗳️ 3 Consensus Algorithms: Majority, Weighted (Queen 3x), Byzantine (f < n/3)
- 🧠 Collective Memory: Shared knowledge, LRU cache, SQLite persistence with WAL
- ⚡ Performance: Fast batch spawning with parallel agent coordination
🧠 Intelligence & Memory — How the system learns and remembers
The system stores successful patterns in vector memory, builds a knowledge graph for structural understanding, learns from outcomes via neural networks, and adapts routing based on what works best.
| Layer | Components | What It Does |
|---|---|---|
| Memory | HNSW, AgentDB, Cache | Stores and retrieves patterns with fast HNSW search |
| Knowledge Graph | MemoryGraph, PageRank, Communities | Identifies influential insights, detects clusters (ADR-049) |
| Self-Learning | LearningBridge, SONA, ReasoningBank | Triggers learning from insights, confidence lifecycle (ADR-049) |
| Agent Scopes | AgentMemoryScope, 3-scope dirs | Per-agent isolation + cross-agent knowledge transfer (ADR-049) |
| Embeddings | ONNX Runtime, MiniLM | Local vectors without API calls (faster with ONNX runtime) |
| Learning | SONA, MoE, ReasoningBank | Self-improves from results (sub-millisecond pattern matching) |
| Fine-tuning | MicroLoRA, EWC++ | Lightweight adaptation without full retraining |
⚡ Optimization — How to reduce cost and latency
Skip expensive LLM calls for simple tasks using WebAssembly transforms, and compress tokens to reduce API costs by 30-50%.
| Layer | Components | What It Does |
|---|---|---|
| Agent Booster | WASM, AST analysis | Skips LLM for simple edits (<1ms) |
| Token Optimizer | Compression, Caching | Reduces token usage 30-50% |
🔧 Operations — Background services and integrations
Background daemons handle security audits, performance optimization, and session persistence automatically while you work.
| Layer | Components | What It Does |
|---|---|---|
| Background | Daemon, 12 Workers | Auto-runs audits, optimization, learning |
| Security | AIDefence, Validation | Blocks injection, detects threats |
| Sessions | Persist, Restore, Export | Saves context across conversations |
| GitHub | PR, Issues, Workflows | Manages repos and code reviews |
| Analytics | Metrics, Benchmarks | Monitors performance, finds bottlenecks |
🎯 Task Routing — Extend your Claude Code subscription by 250%
Smart routing skips expensive LLM calls when possible. Simple edits use WASM (free), medium tasks use cheaper models. This can extend your Claude Code usage by 250% or save significantly on direct API costs.
| Complexity | Handler | Speed |
|---|---|---|
| Simple | Agent Booster (WASM) | <1ms |
| Medium | Haiku/Sonnet | ~500ms |
| Complex | Opus + Swarm | 2-5s |
⚡ Agent Booster (WASM) — Skip LLM for simple code transforms
Agent Booster uses WebAssembly to handle simple code transformations without calling the LLM at all. When the hooks system detects a simple task, it routes directly to Agent Booster for instant results.
Supported Transform Intents:
| Intent | What It Does | Example |
|---|---|---|
var-to-const |
Convert var/let to const | var x = 1 → const x = 1 |
add-types |
Add TypeScript type annotations | function foo(x) → function foo(x: string) |
add-error-handling |
Wrap in try/catch | Adds proper error handling |
async-await |
Convert promises to async/await | .then() chains → await |
add-logging |
Add console.log statements | Adds debug logging |
remove-console |
Strip console.* calls | Removes all console statements |
Hook Signals:
When you see these in hook output, the system is telling you how to optimize:
# Agent Booster available - skip LLM entirely
[AGENT_BOOSTER_AVAILABLE] Intent: var-to-const
→ Use Edit tool directly, instant (regex-based, no LLM call) than LLM
# Model recommendation for Task tool
[TASK_MODEL_RECOMMENDATION] Use model="haiku"
→ Pass model="haiku" to Task tool for cost savings
Performance:
| Metric | Agent Booster | LLM Call |
|---|---|---|
| Latency | <1ms | 2-5s |
| Cost | $0 | $0.0002-$0.015 |
| Speedup | instant (regex-based, no LLM call) | baseline |
💰 Token Optimizer — reduces token usage via pattern caching and smart routing
The Token Optimizer integrates agentic-flow optimizations to reduce API costs by compressing context and caching results.
Savings Breakdown:
| Optimization | Token Savings | How It Works |
|---|---|---|
| ReasoningBank retrieval | -32% | Fetches relevant patterns instead of full context |
| Agent Booster edits | -15% | Simple edits skip LLM entirely |
| Cache (95% hit rate) | -10% | Reuses embeddings and patterns |
| Optimal batch size | -20% | Groups related operations |
| Combined | 30-50% | Stacks multiplicatively |
Usage:
import { getTokenOptimizer } from '@claude-flow/integration';
const optimizer = await getTokenOptimizer();
// Get compact context (32% fewer tokens)
const ctx = await optimizer.getCompactContext("auth patterns");
// Optimized edit (instant (regex-based, no LLM call) for simple transforms)
await optimizer.optimizedEdit(file, oldStr, newStr, "typescript");
// Optimal config for swarm (100% success rate)
const config = optimizer.getOptimalConfig(agentCount);
🛡️ Anti-Drift Swarm Configuration — Prevent goal drift in multi-agent work
Complex swarms can drift from their original goals. Ruflo V3 includes anti-drift defaults that prevent agents from going off-task.
Recommended Configuration:
// Anti-drift defaults (ALWAYS use for coding tasks)
swarm_init({
topology: "hierarchical", // Single coordinator enforces alignment
maxAgents: 8, // Smaller team = less drift surface
strategy: "specialized" // Clear roles reduce ambiguity
})
Why This Prevents Drift:
| Setting | Anti-Drift Benefit |
|---|---|
hierarchical |
Coordinator validates each output against goal, catches divergence early |
maxAgents: 6-8 |
Fewer agents = less coordination overhead, easier alignment |
specialized |
Clear boundaries - each agent knows exactly what to do, no overlap |
raft consensus |
Leader maintains authoritative state, no conflicting decisions |
Additional Anti-Drift Measures:
- Frequent checkpoints via
post-taskhooks - Shared memory namespace for all agents
- Short task cycles with verification gates
- Hierarchical coordinator reviews all outputs
Task → Agent Routing (Anti-Drift):
| Code | Task Type | Recommended Agents |
|---|---|---|
| 1 | Bug Fix | coordinator, researcher, coder, tester |
| 3 | Feature | coordinator, architect, coder, tester, reviewer |
| 5 | Refactor | coordinator, architect, coder, reviewer |
| 7 | Performance | coordinator, perf-engineer, coder |
| 9 | Security | coordinator, security-architect, auditor |
| 11 | Memory | coordinator, memory-specialist, perf-engineer |
Claude Code: With vs Without Ruflo
| Capability | Claude Code Alone | Claude Code + Ruflo |
|---|---|---|
| Agent Collaboration | Agents work in isolation, no shared context | Agents collaborate via swarms with shared memory and consensus |
| Coordination | Manual orchestration between tasks | Queen-led hierarchy with 3 consensus algorithms (Raft, Byzantine, Gossip) |
| Hive Mind | ⛔ Not available | 🐝 Queen-led swarms with collective intelligence, 3 queen types, 8 worker types |
| Consensus | ⛔ No multi-agent decisions | Byzantine fault-tolerant voting (f < n/3), weighted, majority |
| Memory | Session-only, no persistence | HNSW vector memory with sub-ms retrieval + knowledge graph |
| Vector Database | ⛔ No native support | 🐘 RuVector PostgreSQL with 77+ SQL functions, ~61µs search, 16,400 QPS |
| Knowledge Graph | ⛔ Flat insight lists | PageRank + community detection identifies influential insights (ADR-049) |
| Collective Memory | ⛔ No shared knowledge | Shared knowledge base with LRU cache, SQLite persistence, 8 memory types |
| Learning | Static behavior, no adaptation | SONA self-learning with sub-millisecond pattern matching, LearningBridge for insights |
| Agent Scoping | Single project scope | 3-scope agent memory (project/local/user) with cross-agent transfer |
| Task Routing | You decide which agent to use | Intelligent routing based on learned patterns (89% accuracy) |
| Complex Tasks | Manual breakdown required | Automatic decomposition across 5 domains (Security, Core, Integration, Support) |
| Background Workers | Nothing runs automatically | 12 context-triggered workers auto-dispatch on file changes, patterns, sessions |
| LLM Provider | Anthropic only | 5 providers (Anthropic, OpenAI, Google, Cohere, Ollama) with automatic failover and cost-based routing (cost-optimized routing) |
| Security | Standard protections | CVE-hardened with bcrypt, input validation, path traversal prevention |
| Performance | Baseline | Faster tasks via parallel swarm spawning and intelligent routing |
Quick Start
Prerequisites
- Node.js 20+ (required)
- npm 9+ / pnpm / bun package manager
IMPORTANT: Claude Code must be installed first:
# 1. Install Claude Code globally
npm install -g @anthropic-ai/claude-code
# 2. (Optional) Skip permissions check for faster setup
claude --dangerously-skip-permissions
Installation
One-Line Install (Recommended)
# curl-style installer with progress display
curl -fsSL https://cdn.jsdelivr.net/gh/ruvnet/ruflo@main/scripts/install.sh | bash
# Full setup (global + MCP + diagnostics)
curl -fsSL https://cdn.jsdelivr.net/gh/ruvnet/ruflo@main/scripts/install.sh | bash -s -- --full
Install Options
| Option | Description |
|---|---|
--global, -g |
Install globally (npm install -g) |
--minimal, -m |
Skip optional deps (faster, ~15s) |
--setup-mcp |
Auto-configure MCP server for Claude Code |
--doctor, -d |
Run diagnostics after install |
--no-init |
Skip project initialization (init runs by default) |
--full, -f |
Full setup: global + MCP + doctor |
--version=X.X.X |
Install specific version |
Examples:
# Minimal global install (fastest)
curl ... | bash -s -- --global --minimal
# With MCP auto-setup
curl ... | bash -s -- --global --setup-mcp
# Full setup with diagnostics
curl ... | bash -s -- --full
Speed:
| Mode | Time |
|---|---|
| npx (cached) | ~3s |
| npx (fresh) | ~20s |
| global | ~35s |
| --minimal | ~15s |
npm/npx Install
# Quick start (no install needed)
npx ruflo@latest init
# Or install globally
npm install -g ruflo@latest
ruflo init
# With Bun (faster)
bunx ruflo@latest init
Install Profiles
| Profile | Size | Use Case |
|---|---|---|
--omit=optional |
~45MB | Core CLI only (fastest) |
| Default | ~340MB | Full install with ML/embeddings |
# Minimal install (skip ML/embeddings)
npm install -g ruflo@latest --omit=optional
🤖 OpenAI Codex CLI Support — Full Codex integration with self-learning
Ruflo supports both Claude Code and OpenAI Codex CLI via the @claude-flow/codex package, following the Agentics Foundation standard.
Quick Start for Codex
# Initialize for Codex CLI (creates AGENTS.md instead of CLAUDE.md)
npx ruflo@latest init --codex
# Full Codex setup with all 137+ skills
npx ruflo@latest init --codex --full
# Initialize for both platforms (dual mode)
npx ruflo@latest init --dual
Platform Comparison
| Feature | Claude Code | OpenAI Codex |
|---|---|---|
| Config File | CLAUDE.md |
AGENTS.md |
| Skills Dir | .claude/skills/ |
.agents/skills/ |
| Skill Syntax | /skill-name |
$skill-name |
| Settings | settings.json |
config.toml |
| MCP | Native | Via codex mcp add |
| Default Model | claude-sonnet | gpt-5.3 |
Key Concept: Execution Model
┌─────────────────────────────────────────────────────────────────┐
│ CLAUDE-FLOW = ORCHESTRATOR (tracks state, stores memory) │
│ CODEX = EXECUTOR (writes code, runs commands, implements) │
└─────────────────────────────────────────────────────────────────┘
Codex does the work. Claude-flow coordinates and learns.
Dual-Mode Integration (Claude Code + Codex)
Run Claude Code for interactive development and spawn headless Codex workers for parallel background tasks:
┌─────────────────────────────────────────────────────────────────┐
│ CLAUDE CODE (interactive) ←→ CODEX WORKERS (headless) │
│ - Main conversation - Parallel background execution │
│ - Complex reasoning - Bulk code generation │
│ - Architecture decisions - Test execution │
│ - Final integration - File processing │
└─────────────────────────────────────────────────────────────────┘
# Spawn parallel Codex workers from Claude Code
claude -p "Analyze src/auth/ for security issues" --session-id "task-1" &
claude -p "Write unit tests for src/api/" --session-id "task-2" &
claude -p "Optimize database queries in src/db/" --session-id "task-3" &
wait # Wait for all to complete
| Dual-Mode Feature | Benefit |
|---|---|
| Parallel Execution | 4-8x faster for bulk tasks |
| Cost Optimization | Route simple tasks to cheaper workers |
| Context Preservation | Shared memory across platforms |
| Best of Both | Interactive + batch processing |
Dual-Mode CLI Commands (NEW)
# List collaboration templates
npx @claude-flow/codex dual templates
# Run feature development swarm (architect → coder → tester → reviewer)
npx @claude-flow/codex dual run --template feature --task "Add user auth"
# Run security audit swarm (scanner → analyzer → fixer)
npx @claude-flow/codex dual run --template security --task "src/auth/"
# Run refactoring swarm (analyzer → planner → refactorer → validator)
npx @claude-flow/codex dual run --template refactor --task "src/legacy/"
Pre-Built Collaboration Templates
| Template | Pipeline | Platforms |
|---|---|---|
| feature | architect → coder → tester → reviewer | Claude + Codex |
| security | scanner → analyzer → fixer | Codex + Claude |
| refactor | analyzer → planner → refactorer → validator | Claude + Codex |
MCP Integration for Codex
When you run init --codex, the MCP server is automatically registered:
# Verify MCP is registered
codex mcp list
# If not present, add manually:
codex mcp add ruflo -- npx ruflo mcp start
Self-Learning Workflow
1. LEARN: memory_search(query="task keywords") → Find similar patterns
2. COORD: swarm_init(topology="hierarchical") → Set up coordination
3. EXECUTE: YOU write code, run commands → Codex does real work
4. REMEMBER: memory_store(key, value, namespace="patterns") → Save for future
The Intelligence Loop (ADR-050) automates this cycle through hooks. Each session automatically:
- Builds a knowledge graph from memory entries (PageRank + Jaccard similarity)
- Injects ranked context into every route decision
- Tracks edit patterns and generates new insights
- Boosts confidence for useful patterns, decays unused ones
- Saves snapshots so you can track improvement with
node .claude/helpers/hook-handler.cjs stats
MCP Tools for Learning
| Tool | Purpose | When to Use |
|---|---|---|
memory_search |
Semantic vector search | BEFORE starting any task |
memory_store |
Save patterns with embeddings | AFTER completing successfully |
swarm_init |
Initialize coordination | Start of complex tasks |
agent_spawn |
Register agent roles | Multi-agent workflows |
neural_train |
Train on patterns | Periodic improvement |
137+ Skills Available
| Category | Examples |
|---|---|
| V3 Core | $v3-security-overhaul, $v3-memory-unification, $v3-performance-optimization |
| AgentDB | $agentdb-vector-search, $agentdb-optimization, $agentdb-learning |
| Swarm | $swarm-orchestration, $swarm-advanced, $hive-mind-advanced |
| GitHub | $github-code-review, $github-workflow-automation, $github-multi-repo |
| SPARC | $sparc-methodology, $sparc:architect, $sparc:coder, $sparc:tester |
| Flow Nexus | $flow-nexus-neural, $flow-nexus-swarm, $flow-nexus:workflow |
| Dual-Mode | $dual-spawn, $dual-coordinate, $dual-collect |
Vector Search Details
- Embedding Dimensions: 384
- Search Algorithm: HNSW (sub-millisecond)
- Similarity Scoring: 0-1 (higher = better)
- Score > 0.7: Strong match, use pattern
- Score 0.5-0.7: Partial match, adapt
- Score < 0.5: Weak match, create new
Basic Usage
# Initialize project
npx ruflo@latest init
# Start MCP server for Claude Code integration
npx ruflo@latest mcp start
# Spawn a coding agent
npx ruflo@latest agent spawn -t coder --name my-coder
# Launch a hive-mind swarm with an objective
npx ruflo@latest hive-mind spawn "Implement user authentication"
# List available agent types
npx ruflo@latest agent list
Upgrading
# Update helpers and statusline (preserves your data)
npx ruflo@latest init upgrade
# Update AND add any missing skills/agents/commands
npx ruflo@latest init upgrade --add-missing
The --add-missing flag automatically detects and installs new skills, agents, and commands that were added in newer versions, without overwriting your existing customizations.
Claude Code MCP Integration
Add ruflo as an MCP server for seamless integration:
# Add ruflo MCP server to Claude Code
claude mcp add ruflo -- npx -y ruflo@latest mcp start
# Verify installation
claude mcp list
Once added, Claude Code can use all 313 ruflo MCP tools directly:
swarm_init- Initialize agent swarmsagent_spawn- Spawn specialized agentsmemory_search- Search patterns with HNSW vector searchhooks_route- Intelligent task routing- And 255+ more tools...
What is it exactly? Agents that learn, build and work perpetually.
🆚 Why Ruflo v3?
Ruflo v3 introduces self-learning neural capabilities that no other agent orchestration framework offers. While competitors require manual agent configuration and static routing, Ruflo learns from every task execution, prevents catastrophic f
Similar mcp servers
claude-task-master
MCP server for AI task management — streamlines development workflows with Claude and other AI chats
n8n-mcp
MCP server for n8n workflow automation — provides AI assistants deep access to 1,500+ n8n nodes and templates
ccusage
MCP server for Claude Code usage — real-time token and cost analysis from local JSONL files
container-use
MCP server for coding agents — provides isolated, ephemeral container environments for parallel development