Zum Inhalt springen

MCP Server / MassGen

MassGen

πŸš€ MassGen is an open-source multi-agent scaling system that runs in your terminal, autonomously orchestrating frontier models and agents to collaborate, reason, and produce high-quality results. | Join us on Discord: discord.massgen.ai

β˜… 941von @massgenNOASSERTIONGitHub β†’

Installation

Claude Code
claude mcp add massgen -- npx -y skills
npx
npx -y skills

npm: skills

Transport

stdiossehttp

Tools (20)

Feature

Description

medium

high`; Codex GPT-5 models also include `xhigh`) - Auto-detect Docker availability and configure execution mode - If Docker mode is selected, show a Skills step where you can choose package(s) (`opensk

Description

mcp_servers

dict

string

Yes

string

stdio only

list

stdio only

string

http only

dict

No

allowed_tools

list

exclude_tools

list

Description

cwd

string

snapshot_storage

string

agent_temporary_workspace

string

Description

context_paths

list

string

Yes

string

Yes

list

No

Dokumentation

MassGen is a cutting-edge multi-agent framework that coordinates AI agents to solve complex tasks through redundancy and iterative refinement. Every agent tackles the full problem, observing, critiquing, and building on each other's work across cycles of refinement and restarts. When agents believe there is a strong enough answer, they vote, and the best collectively validated answer wins. This approach to parallel refinement and collective validation lays the groundwork for principled multi-agent scaling, where the system continuously improves its outputs by leveraging diverse agent perspectives and enforcing quality through consensus.

This project started with the "threads of thought" and "iterative refinement" ideas presented in The Myth of Reasoning, and extends the classic "multi-agent conversation" idea in AG2. Here is a video recording of the background context introduction presented at the Berkeley Agentic AI Summit 2025.


πŸ“‹ Table of Contents


✨ Key Features

| Feature | Description | |---------|-------------| | 🀝 Cross-Model/Agent Synergy | Harness strengths from diverse frontier model-powered agents | | ⚑ Parallel Processing | Multiple agents tackle problems simultaneously | | πŸ‘₯ Intelligence Sharing | Agents share and learn from each other's work | | πŸ”„ Consensus Building | Natural convergence through collaborative refinement | | πŸ–₯️ Live Visualization | Interactive Textual TUI with timeline, agent cards, and vote tracking (default). Also available: Web UI, Rich display. |


πŸ†• Latest Features (v0.1.75)

πŸŽ‰ Released: April 10, 2026

What's New in v0.1.75:

  • πŸͺ Codex Native Hooks - Hybrid hook system for Codex backend combining native and MCP capabilities.
  • πŸ›‘οΈ Checkpoint WebUI Auto-Launch - Checkpoint workflows auto-launch the WebUI for visual monitoring.
  • πŸ“– Standalone MCP Server Docs - Guide for massgen-checkpoint-mcp with safety policy integration.

Try v0.1.75 Features:

pip install massgen==0.1.75
uv run massgen --config @examples/features/fast_iteration.yaml "Create an svg of an AI agent coding."

β†’ See full release history and examples


πŸ—οΈ System Design

MassGen operates through an architecture designed for seamless multi-agent collaboration:

graph TB
    O[πŸš€ MassGen Orchestrator<br/>πŸ“‹ Task Distribution & Coordination]

    subgraph Collaborative Agents
        A1[Agent 1<br/>πŸ—οΈ Anthropic/Claude + Tools]
        A2[Agent 2<br/>🌟 Google/Gemini + Tools]
        A3[Agent 3<br/>πŸ€– OpenAI/GPT + Tools]
        A4[Agent 4<br/>⚑ xAI/Grok + Tools]
    end

    H[πŸ”„ Shared Collaboration Hub<br/>πŸ“‘ Real-time Notification & Consensus]

    O --> A1 & A2 & A3 & A4
    A1 & A2 & A3 & A4 <--> H

    classDef orchestrator fill:#e1f5fe,stroke:#0288d1,stroke-width:3px
    classDef agent fill:#f3e5f5,stroke:#7b1fa2,stroke-width:2px
    classDef hub fill:#e8f5e8,stroke:#388e3c,stroke-width:2px

    class O orchestrator
    class A1,A2,A3,A4 agent
    class H hub

The system's workflow is defined by the following key principles:

Parallel Processing - Multiple agents tackle the same task simultaneously, each leveraging their unique capabilities (different models, tools, and specialized approaches).

Real-time Collaboration - Agents continuously share their working summaries and insights through a notification system, allowing them to learn from each other's approaches and build upon collective knowledge.

Convergence Detection - The system intelligently monitors when agents have reached stability in their solutions and achieved consensus through natural collaboration rather than forced agreement.

Adaptive Coordination - Agents can restart and refine their work when they receive new insights from others, creating a dynamic and responsive problem-solving environment.

This collaborative approach ensures that the final output leverages collective intelligence from multiple AI systems, leading to more robust and well-rounded results than any single agent could achieve alone.


πŸ“– Complete Documentation: For comprehensive guides, API reference, and detailed examples, visit MassGen Official Documentation


πŸš€ Quick Start

1. πŸ“₯ Installation

Method 1: PyPI Installation (Recommended - Python 3.11+):

# Install MassGen via pip
pip install massgen

# Or with uv (faster)
pip install uv
uv venv && source .venv/bin/activate
uv pip install massgen

# If you install massgen in uv, make sure you either activate your venv using source .venv/bin/activate
# Or include "uv run" before all commands

Quickstart Setup (Fastest way to get running):

# Step 1: Set up API keys, Docker, and skills
uv run massgen --setup

# Step 2: Create a simple config and start
uv run massgen --quickstart

The --setup command will:

  • Configure your API keys (OpenAI, Anthropic, Google, xAI)
  • Offer to set up Docker images for code execution
  • Offer to install skills (openskills, Anthropic/OpenAI/Vercel collections, Agent Browser skill, Crawl4AI)

The --quickstart command will:

  • Ask how many agents you want (1-5, default 3)
  • Ask which backend/model for each agent
  • For GPT-5x models, ask for reasoning.effort (low|medium|high; Codex GPT-5 models also include xhigh)
  • Auto-detect Docker availability and configure execution mode
  • If Docker mode is selected, show a Skills step where you can choose package(s) (openskills-based Anthropic/OpenAI/Vercel/Agent Browser plus Crawl4AI) and install them in-place with live status
  • Create a ready-to-use config and launch into interactive TUI mode

πŸ€– Use MassGen from Your AI Coding Agent:

Install the MassGen skill to invoke MassGen directly from Claude Code, OpenAI Codex, GitHub Copilot, Cursor, and 40+ other agents that support the Agent Skills standard:

npx skills add massgen/skills

Then use /massgen (Claude Code) or $massgen (Codex) to run multi-agent evaluation, planning, spec writing, or any general task. See the skills docs for per-agent install options.

πŸ–₯️ Textual TUI (Default Display Mode):

MassGen launches with an interactive Terminal User Interface (TUI) by default, providing:

  • πŸ“Š Real-time timeline of all agent activities
  • 🎯 Individual agent status cards for each team member
  • πŸ—³οΈ Vote visualization and consensus tracking
  • πŸ’¬ Multi-turn conversation management
  • ⌨️ Keyboard controls for navigation (↑/↓ to scroll, 'q' to cancel)

Legacy Rich display:

massgen --display rich "Your question"

Alternative: Full Setup Wizard

For more control, use the full configuration wizard:

uv run massgen --init

This guides you through use case selection (Research, Code, Q&A, etc.) and advanced configuration options.

After setup:

# Interactive mode
uv run massgen

# Single query
uv run massgen "Your question here"

# With example configurations
uv run massgen --config @examples/basic/multi/three_agents_default "Your question"

β†’ See Installation Guide for complete setup instructions.

Method 2: Development Installation (for contributors):

Clone the repository

git clone https://github.com/Leezekun/MassGen.git
cd MassGen

Install in editable mode with pip

Option 1 (recommended): Installing with uv (faster)

uv venv
source .venv/bin/activate  # Windows: .venv\Scripts\activate
uv pip install -e .

# If you install massgen in uv, make sure you either activate your venv using source .venv/bin/activate
# Or include "uv run" before all commands

# Automated setup (works on all platforms) - installs dependencies, skills, Docker images, also sets up API keys
uv run massgen --setup

# Or use the bash script (Unix/Linux/macOS only), need manually config API keys, see sections below
uv run ./scripts/init.sh

# If you would like to install other dependencies later
# Here is a light-weighted setup script which only installs skills (works on all platforms)
uv run massgen --setup-skills

# Or use the bash script (Unix/Linux/macOS only)
uv run ./scripts/init_skills.sh

Option 2: Using traditional Python env

pip install -e .

# Optional: External framework integration
pip install -e ".[external]"

# Automated setup (works on all platforms) - installs dependencies, skills, Docker images, also sets up API keys
massgen --setup

# Or use the bash script (Unix/Linux/macOS only), need manually config API keys, see sections below
./scripts/init.sh

# If you would like to install other dependencies later
# Here is a light-weighted setup script which only installs skills (works on all platforms)
massgen --setup-skills

# Or use the bash script (Unix/Linux/macOS only)
./scripts/init_skills.sh

Note: The --setup and --setup-skills commands work cross-platform (Windows, macOS, Linux). The bash scripts (init.sh, init_skills.sh) are Unix-only but provide additional dev setup like Docker image builds.

Using uv with venv:

git clone https://github.com/Leezekun/MassGen.git
cd MassGen
uv venv
source .venv/bin/activate  # Windows: .venv\Scripts\activate
uv pip install -e .

Using traditional Python venv:

git clone https://github.com/Leezekun/MassGen.git
cd MassGen
python -m venv .venv
source .venv/bin/activate  # Windows: .venv\Scripts\activate
pip install -e .

Global installation with uv tool:

git clone https://github.com/Leezekun/MassGen.git
cd MassGen
uv tool install -e .
# Now run from any directory
uv tool run massgen --config @examples/basic/multi/three_agents_default "Question"

Backwards compatibility (uv run):

cd /path/to/MassGen
uv run massgen --config @examples/basic/multi/three_agents_default "Question"
uv run python -m massgen.cli --config config.yaml "Question"

Optional CLI Tools:

# Claude Code CLI - Advanced coding assistant
npm install -g @anthropic-ai/claude-code

# LM Studio - Local model inference
# MacOS/Linux:
sudo ~/.lmstudio/bin/lms bootstrap
# Windows:
cmd /c %USERPROFILE%\.lmstudio\bin\lms.exe bootstrap

After setup:

# Interactive mode
uv run massgen

# Single query
uv run massgen "Your question here"

# With example configurations
uv run massgen --config @examples/basic/multi/three_agents_default "Your question"

2. πŸ” API Configuration

Create a .env file in your working directory with your API keys:

# Copy this template to .env and add your API keys
OPENAI_API_KEY=sk-...
ANTHROPIC_API_KEY=sk-ant-...
GOOGLE_API_KEY=...
XAI_API_KEY=...

# Optional: Additional providers
CEREBRAS_API_KEY=...
TOGETHER_API_KEY=...
GROQ_API_KEY=...
OPENROUTER_API_KEY=...

MassGen automatically loads API keys from .env in your current directory.

β†’ Complete setup guide with all providers: See API Key Configuration in the docs

Get API keys:

3. 🧩 Supported Models and Tools

Models

The system currently supports multiple model providers with advanced capabilities:

API-based Models:

  • OpenAI: GPT-5.2 (recommended default), GPT-5.1, GPT-5 series (GPT-5, GPT-5-mini, GPT-5-nano), GPT-5.1-Codex series, GPT-4.1 series, GPT-4o, o4-mini with reasoning, web search, code interpreter, and computer-use support
    • Note: We recommend GPT-5.2/5.1/5 over Codex models. Codex models are optimized for shorter system messages and may not work well with MassGen's coordination prompts.
    • Reasoning: GPT-5.1 and GPT-5.2 default to reasoning: none. MassGen automatically sets reasoning.effort: medium when no reasoning config is provided, matching GPT-5's default behavior.
  • Azure OpenAI: Any Azure-deployed models (GPT-4, GPT-4o, GPT-35-turbo, etc.)
  • Claude / Anthropic: Claude Opus 4.5, Claude Haiku 4.5, Claude Sonnet 4.5, Claude Opus 4.1, Claude Sonnet 4
    • Advanced tooling: web search, code execution, Files API, programmatic tool calling, tool search with deferred loading
  • Claude Code: Native Claude Code SDK with server-side session persistence and built-in dev tools
  • Gemini: Gemini 3 Pro, Gemini 2.5 Flash, Gemini 2.5 Pro with code execution and grounding
  • Grok / xAI: Grok-4.1, Grok-4, Grok-3, Grok-3-mini with Grok Live Search
  • Cerebras AI: Ultra-fast inference for supported models
  • Together AI, Fireworks AI, Groq: Fast inference for LLaMA, Mistral, Qwen, and other open models
  • OpenRouter: Multi-model aggregator with dynamic model listing (400+ models)
  • Kimi / Moonshot: Chinese AI models via OpenAI-compatible API
  • Nebius AI Studio: Cloud inference platform
  • POE: Quora AI platform with dynamic model discovery
  • Qwen / Alibaba: DashScope API for Qwen models
  • Z AI / Zhipu: GLM-4.5 and related models

Local Model Support:

  • vLLM & SGLang: Unified inference backend supporting both vLLM and SGLang servers

    • vLLM (port 8000) and SGLang (port 30000) with OpenAI-compatible API
    • Support for top_k, repetition_penalty, chat_template_kwargs parameters
    • SGLang-specific separate_reasoning parameter for thinking models
    • Mixed server deployments with configuration example: two_qwen_vllm_sglang.yaml
  • LM Studio: Run open-weight models locally with automatic server management

    • Automatic LM Studio CLI installation
    • Auto-download and loading of models
    • Support for LLaMA, Mistral, Qwen and other open-weight models

β†’ For complete model list and configuration details, see Supported Models

Tools

MassGen agents can leverage various tools to enhance their problem-solving capabilities:

  • Built-in Tools: Web search, code execution, bash/shell (provider-dependent)
  • Filesystem: Native file operations or via MCP
  • MCP Integration: Connect to any MCP server for extended capabilities
  • Custom Tools: Define your own tools via YAML configuration
  • Multimodal: Image, audio, video understanding and generation (native or via custom tools)

β†’ For detailed backend capabilities and tool support matrix, see User Guide - Backends


4. πŸƒ Run MassGen

Complete Usage Guide: For all usage modes, advanced features, and interactive multi-turn sessions, see Running MassGen

πŸš€ Getting Started

CLI Configuration Parameters

| Parameter | Description | |-------------------|-------------| | --config | Path to YAML configuration file with agent definitions, model parameters, backend parameters and UI settings | | --backend | Backend type for quick setup without a config file (claude, claude_code, gemini, grok, openai, azure_openai, zai). Optional for models with default backends.| | --model | Model name for quick setup (e.g., gemini-2.5-flash, gpt-5-nano, ...). --config and --model are mutually exclusive - use one or the other. | | --system-message | System prompt for the agent in quick setup mode. If --config is provided, --system-message is omitted. | | --cwd-context | Add current working directory as runtime context path: ro/read for read-only, rw/write for write access. In TUI, this initializes the same state as Ctrl+P. | | --plan | Planning-only mode. Agents create a structured task plan without auto-executing it. | | --plan-depth | Plan granularity for --plan: dynamic, shallow, medium, or deep. | | --plan-and-execute | Run both phases: create a plan, then execute it automatically. | | --execute-plan | Execute an existing plan by path, plan ID, or latest. | | --no-display | Disable real-time streaming UI coordination display (fallback to simple text output).| | --no-logs | Disable real-time logging.| | --debug | Enable debug mode with verbose logging (NEW in v0.0.13). Shows detailed orchestrator activities, agent messages, backend operations, and tool calls. Debug logs are saved to agent_outputs/log_{time}/massgen_debug.log. | | "<your question>" | Optional single-question input; if omitted, MassGen enters interactive chat mode. |

0. OpenAI-Compatible HTTP Server (NEW)

Run MassGen as an OpenAI-compatible HTTP API (FastAPI + Uvicorn). This is useful for integrating MassGen with existing tooling that expects POST /v1/chat/completions.

# Start server (defaults: host 0.0.0.0, port 4000)
massgen serve

# With explicit bind + defaults for model/config
massgen serve --host 0.0.0.0 --port 4000 --config path/to/config.yaml --default-model gpt-5

Endpoints

  • GET /health
  • POST /v1/chat/completions (supports stream: true SSE and OpenAI-style tool calling)

cURL examples

# Health
curl http://localhost:4000/health

# Non-streaming chat completion
curl http://localhost:4000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "massgen",
    "messages": [{"role": "user", "content": "hi"}],
    "stream": false
  }'

# Streaming (Server-Sent Events)
curl -N http://localhost:4000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "massgen",
    "messages": [{"role": "user", "content": "hi"}],
    "stream": true
  }'

Notes

  • Client-provided tools are supported, but tool names that collide with MassGen workflow tools are rejected.
  • Environment variables (optional): MASSGEN_SERVER_HOST, MASSGEN_SERVER_PORT, MASSGEN_SERVER_DEFAULT_CONFIG, MASSGEN_SERVER_DEFAULT_MODEL, MASSGEN_SERVER_DEBUG.

1. Single Agent (Easiest Start)

Quick Start Commands:

# Quick test with any supported model - no configuration needed
uv run python -m massgen.cli --model claude-sonnet-4-5-20250929 "What is machine learning?"
uv run python -m massgen.cli --model gemini-3-pro-preview "Explain quantum computing"
uv run python -m massgen.cli --model gpt-5-nano "Summarize the latest AI developments"

Configuration:

Use the agent field to define a single agent with its backend and settings:

agent:
  id: "<agent_name>"
  backend:
    type: "azure_openai" | "chatcompletion" | "claude" | "claude_code" | "gemini" | "grok" | "openai" | "zai" | "lmstudio" #Type of backend
    model: "<model_name>" # Model name
    api_key: "<optional_key>"  # API key for backend. Uses env vars by default.
  system_message: "..."    # System Message for Single Agent

β†’ See all single agent configs

2. Multi-Agent Collaboration (Recommended)

Configuration:

Use the agents field to define multiple agents, each with its own backend and config:

Quick Start Commands:

# Three powerful agents working together - Gemini, GPT-5, and Grok
massgen --config @examples/basic/multi/three_agents_default \
  "Analyze the pros and cons of renewable energy"

This showcases MassGen's core strength:

  • Gemini 3 Pro - Fast research with web search
  • GPT-5 Nano - Advanced reasoning with code execution
  • Grok-4 Fast - Real-time information and alternative perspectives
agents:  # Multiple agents (alternative to 'agent')
  - id: "<agent1 name>"
    backend:
      type: "azure_openai" | "chatcompletion" | "claude" | "claude_code" | "gemini" | "grok" | "openai" |  "zai" | "lmstudio" #Type of backend
      model: "<model_name>" # Model name
      api_key: "<optional_key>"  # API key for backend. Uses env vars by default.
    system_message: "..."    # System Message for Single Agent
  - id: "..."
    backend:
      type: "..."
      model: "..."
      ...
    system_message: "..."

β†’ Explore more multi-agent setups

3. Model context protocol (MCP)

The Model context protocol (MCP) standardises how applications expose tools and context to language models. From the official documentation:

MCP is an open protocol that standardizes how applications provide context to LLMs. Think of MCP like a USB-C port for AI applications. Just as USB-C provides a standardized way to connect your devices to various peripherals and accessories, MCP provides a standardized way to connect AI models to different data sources and tools.

MCP Configuration Parameters:

| Parameter | Type | Required | Description | |-----------|------|----------|-------------| | mcp_servers | dict | Yes (for MCP) | Container for MCP server definitions | | └─ type | string | Yes | Transport: "stdio" or "streamable-http" | | └─ command | string | stdio only | Command to run the MCP server | | └─ args | list | stdio only | Arguments for the command | | └─ url | string | http only | Server endpoint URL | | └─ env | dict | No | Environment variables to pass | | allowed_tools | list | No | Whitelist specific tools (if omitted, all tools available) | | exclude_tools | list | No | Blacklist dangerous/unwanted tools |

Quick Start Commands (Check backend MCP support here):

# Weather service with GPT-5
massgen --config @examples/tools/mcp/gpt5_nano_mcp_example \
  "What's the weather forecast for New York this week?"

# Multi-tool MCP with Gemini - Search + Weather + Filesystem (Requires BRAVE_API_KEY in .env)
massgen --config @examples/tools/mcp/multimcp_gemini \
  "Find the best restaurants in Paris and save the recommendations to a file"

Configuration:

agents:
  # Basic MCP Configuration:
  backend:
    type: "openai"              # Your backend choice
    model: "gpt-5-mini"         # Your model choice

    # Add MCP servers here
    mcp_servers:
      weather:                  # Server name (you choose this)
        type: "stdio"           # Communication type
        command: "npx"          # Command to run
        args: ["-y", "@modelcontextprotocol/server-weather"]  # MCP server package

  # That's it! The agent can now check weather.

  # Multiple MCP Tools Example:
  backend:
    type: "gemini"
    model: "gemini-3.0-pro-preview"
    mcp_servers:
      # Web search
      search:
        type: "stdio"
        command: "npx"
        args: ["-y", "@modelcontextprotocol/server-brave-search"]
        env:
          BRAVE_API_KEY: "${BRAVE_API_KEY}"  # Set in .env file

      # HTTP-based MCP server (streamable-http transport)
      custodm_api:
        type: "streamable-http"   # For HTTP/SSE servers
        url: "http://localhost:8080/mcp/sse"  # Server endpoint


  # Tool configuration (MCP tools are auto-discovered)
  allowed_tools:                        # Optional: whitelist specific tools
    - "mcp__weather__get_current_weather"
    - "mcp__test_server__mcp_echo"
    - "mcp__test_server__add_numbers"

  exclude_tools:                        # Optional: blacklist specific tools
    - "mcp__test_server__current_time"

β†’ View more MCP examples

β†’ For comprehensive MCP integration guide, see MCP Integration

4. File System Operations & Workspace Management

MassGen provides comprehensive file system support through multiple backends, enabling agents to read, write, and manipulate files in organized workspaces.

Filesystem Configuration Parameters:

| Parameter | Type | Required | Description | |-----------|------|----------|-------------| | cwd | string | Yes (for file ops) | Working directory for file operations (agent-specific workspace) | | snapshot_storage | string | Yes | Directory for workspace snapshots | | agent_temporary_workspace | string | Yes | Parent directory for temporary workspaces |

Quick Start Commands:

# File operations with Claude Code
massgen --config @examples/tools/filesystem/claude_code_single \
  "Create a Python web scraper and save results to CSV"

# Multi-agent file collaboration
massgen --config @examples/tools/filesystem/claude_code_context_sharing \
  "Generate a comprehensive project report with charts and analysis"

Configuration:

# Basic Workspace Setup:
agents:
  - id: "file-agent"
    backend:
      type: "claude_code"          # Backend with file support
      cwd: "workspace"             # Isolated workspace for file operations

# Multi-Agent Workspace Isolation:
agents:
  - id: "agent_a"
    backend:
      type: "claude_code"
      cwd: "workspace1"            # Agent-specific workspace

  - id: "agent_b"
    backend:
      type: "gemini"
      cwd: "workspace2"            # Separate workspace

orchestrator:
  snapshot_storage: "snapshots"              # Shared snapshots directory
  agent_temporary_workspace: "temp_workspaces" # Temporary workspace management

Available File Operations:

  • Claude Code: Built-in tools (Read, Write, Edit, MultiEdit, Bash, Grep, Glob, LS, TodoWrite)
  • Other Backends: Via MCP Filesystem Server

Workspace Management:

  • Isolated Workspaces: Each agent's cwd is fully isolated and writable
  • Snapshot Storage: Share workspace context between Claude Code agents
  • Temporary Workspaces: Agents can access previous coordination results

β†’ View more filesystem examples

⚠️ IMPORTANT SAFETY WARNING

MassGen agents can autonomously read, write, modify, and delete files within their permitted directories.

Before running MassGen with filesystem access:

  • Only grant access to directories you're comfortable with agents modifying
  • Use the permission system to restrict write access where needed
  • Consider testing in an isolated directory or virtual environment first
  • Back up important files before granting write access
  • Review the context_paths configuration carefully

The agents will execute file operations without additional confirmation once permissions are granted.

β†’ For comprehensive file operations guide, see File Operations

5. Project Integration & User Context Paths (NEW in v0.0.21)

Work directly with your existing projects! User Context Paths allow you to share specific directories with all agents while maintaining granular permission control. This enables secure multi-agent collaboration on your real codebases, documentation, and data.

MassGen automatically organizes all its working files under a .massgen/ directory in your project root, keeping your project clean and making it easy to exclude MassGen's temporary files from version control.

Project Integration Parameters:

| Parameter | Type | Required | Description | |-----------|------|----------|-------------| | context_paths | list | Yes (for project integration) | Shared directories for all agents | | └─ path | string | Yes | Absolute or relative path to your project directory (must be directory, not file) | | └─ permission | string | Yes | Access level: "read" or "write" (write applies only to final agent) | | └─ protected_paths | list | No | Files/directories immune from modification (relative to context path) |

⚠️ Important Notes:

  • Context paths must point to directories, not individual files
  • Paths can be absolute or relative (resolved against current working directory)
  • Write permissions apply only to the final agent during presentation phase
  • During coordination, all context paths are read-only to protect your files
  • MassGen validates all paths during startup and will show clear error messages for missing paths or file paths

Quick Start Commands:

# Multi-agent collaboration to improve the website in `massgen/configs/resources/v0.0.21-example
massgen --config @examples/tools/filesystem/gpt5mini_cc_fs_context_path "Enhance the website with: 1) A dark/light theme toggle with smooth transitions, 2) An interactive feature that helps users engage with the blog content (your choice - could be search, filtering by topic, reading time estimates, social sharing, reactions, etc.), and 3) Visual polish with CSS animations or transitions that make the site feel more modern and responsive. Use vanilla JavaScript and be creative with the implementation details."

Configuration:

# Basic Project Integration:
agents:
  - id: "code-reviewer"
    backend:
      type: "claude_code"
      cwd: "workspace"             # Agent's isolated work area

orchestrator:
  context_paths:
    - path: "."                    # Current directory (relative path)
      permission: "write"          # Final agent can create/modify files
      protected_paths:             # Optional: files immune from modification
        - ".env"
        - "config.json"
    - path: "/home/user/my-project/src"  # Absolute path example
      permission: "read"           # Agents can analyze your code

# Advanced: Multi-Agent Project Collaboration
agents:
  - id: "analyzer"
    backend:
      type: "gemini"
      cwd: "analysis_workspace"

  - id: "implementer"
    backend:
      type: "claude_code"
      cwd: "implementation_workspace"

orchestrator:
  context_paths:
    - path: "../legacy-app/src"   # Relative path to existing codebase
      permission: "read"           # Read existing codebase
    - path: "../legacy-app/tests"
      permission: "write"          # Final agent can write new tests
      protected_paths:             # Protect specific test files
        - "integration_tests/production_data_test.py"
    - path: "/home/user/modernized-app"  # Absolute path
      permission: "write"          # Final agent can create modernized version

This showcases project integration:

  • Real Project Access - Work with your actual codebases, not copies
  • Secure Permissions - Granular control over what agents can read/modify
  • Multi-Agent Collaboration - Multiple agents safely work on the same project
  • Context Agents (during coordination): Always READ-only access to protect your files
  • Final Agent (final execution): Gets the configured permission (READ or write)

Use Cases:

  • Code Review: Agents analyze your source code and suggest improvements
  • Documentation: Agents read project docs to understand context and generate updates
  • Data Processing: Agents access shared datasets and generate analysis reports
  • Project Migration: Agents examine existing projects and create modernized versions

Clean Project Organization:

your-project/
β”œβ”€β”€ .massgen/                          # All MassGen state
β”‚   β”œβ”€β”€ sessions/                      # Multi-turn conversation history (if using interactively)
β”‚   β”‚   └── session_20240101_143022/
β”‚   β”‚       β”œβ”€β”€ turn_1/                # Results from turn 1
β”‚   β”‚       β”œβ”€β”€ turn_2/                # Results from turn 2
β”‚   β”‚       └── SESSION_SUMMARY.txt    # Human-readable summary
β”‚   β”œβ”€β”€ workspaces/                    # Agent working directories
β”‚   β”‚   β”œβ”€β”€ agent1/                    # Individual agent workspaces
β”‚   β”‚   └── agent2/
β”‚   β”œβ”€β”€ snapshots/                     # Workspace snapshots for coordination
β”‚   └── temp_workspaces/               # Previous turn results for context
β”œβ”€β”€ massgen/
└── ...

Benefits:

  • βœ… Clean Projects - All MassGen files contained in one directory
  • βœ… Easy Gitignore - Just add .massgen/ to .gitignore
  • βœ… Portable - Move or delete .massgen/ without affecting your project
  • βœ… Multi-Turn Sessions - Conversation history preserved across sessions

Configuration Auto-Organization:

orchestrator:
  # User specifies simple names - MassGen organizes under .massgen/
  snapshot_storage: "snapshots"         # β†’ .massgen/snapshots/
  agent_temporary_workspace: "temp"     # β†’ .massgen/temp/

agents:
  - backend:
      cwd: "workspace1"                 # β†’ .massgen/workspaces/workspace1/

β†’ For comprehensive project integration guide, see Project Integration

Security Considerations:

  • Agent ID Safety: Avoid using agent+incremental digits for IDs (e.g., agent1, agent2). This may cause ID exposure during voting
  • File Access Control: Restrict file access using MCP server configurations when needed
  • Path Validation: All context paths are validated to ensure they exist and are directories (not files)
  • Directory-Only Context Paths: Context paths must point to directories, not individual files

Additional Examples by Provider

Claude (Recursive MCP Execution - v0.0.20+)

# Claude with advanced tool chaining
massgen --config @examples/tools/mcp/claude_mcp_example \
  "Research and compare weather in Beijing and Shanghai"

OpenAI (GPT-5 Series with MCP - v0.0.17+)

# GPT-5 with weather and external tools
massgen --config @examples/tools/mcp/gpt5_nano_mcp_example \
  "What's the weather of Tokyo"

Gemini (Multi-Server MCP - v0.0.15+)

# Gemini with multiple MCP services
massgen --config @examples/tools/mcp/multimcp_gemini \
  "Find accommodations in Paris with neighborhood analysis"    # (requires BRAVE_API_KEY in .env)

Claude Code (Development Tools)

# Professional development environment with auto-configured workspace
uv run python -m massgen.cli \
  --backend claude_code \
  --model sonnet \
  "Create a Flask web app with authentication"

# Default workspace directories created automatically:
# - workspace1/              (working directory)
# - snapshots/              (workspace snapshots)
# - temp_workspaces/        (temporary agent workspaces)

Local Models (LM Studio - v0.0.7+)

# Run open-source models locally
massgen --config @examples/providers/local/lmstudio \
  "Explain machine learning concepts"

β†’ Browse by provider | Browse by tools | Browse teams

Additional Use Case Examples

Question Answering & Research:

# Complex research with multiple perspectives
massgen --config @examples/basic/multi/gemini_gpt5_claude \
  "What's best to do in Stockholm in October 2025"

# Specific research requirements
massgen --config @examples/basic/multi/gemini_gpt5_claude \
  "Give me all the talks on agent frameworks in Berkeley Agentic AI Summit 2025"

Creative Writing:

# Story generation with multiple creative agents
massgen --config @examples/basic/multi/gemini_gpt5_claude \
  "Write a short story about a robot who discovers music"

Development & Coding:

# Full-stack development with file operations
massgen --config @examples/tools/filesystem/claude_code_single \
  "Create a Flask web app with authentication"

Web Automation: (still in test)

# Browser automation with screenshots and reporting
# Prerequisites: npm install @playwright/mcp@latest (for Playwright MCP server)
massgen --config @examples/tools/code-execution/multi_agent_playwright_automation \
  "Browse three issues in https://github.com/Leezekun/MassGen and suggest documentation improvements. Include screenshots and suggestions in a website."

# Data extraction and analysis
massgen --config @examples/tools/code-execution/multi_agent_playwright_automation \
  "Navigate to https://news.ycombinator.com, extract the top 10 stories, and create a summary report"

β†’ See detailed case studies with real session logs and outcomes

Interactive Mode & Advanced Usage

Multi-Turn Conversations:

# Start interactive chat (no initial question)
massgen --config @examples/basic/multi/three_agents_default

# Add CWD context quickly (read-only)
massgen --config @examples/basic/multi/three_agents_default --cwd-context ro

# Add CWD context quickly (read+write)
massgen --config @examples/basic/multi/three_agents_default --cwd-context rw

# Debug mode for troubleshooting
massgen --config @examples/basic/multi/three_agents_default \
  --debug "Your question"

Configuration Files

MassGen configurations are organized by features and use cases. See the Configuration Guide for detailed organization and examples.

Quick navigation:

See MCP server setup guides: Discord MCP | Twitter MCP

Backend Configuration Reference

For detailed configuration of all supported backends (OpenAI, Claude, Gemini, Grok, etc.), see:

β†’ Backend Configuration Guide

Interactive Multi-Turn Mode

MassGen supports an interactive mode where you can have ongoing conversations with the system:

# Start interactive mode with a single agent (no tool enabled by default)
uv run python -m massgen.cli --model gpt-5-mini

# Start interactive mode with configuration file
uv run python -m massgen.cli \
  --config massgen/configs/basic/multi/three_agents_default.yaml

Interactive Mode Features:

  • Multi-turn conversations: Multiple agents collaborate to chat with you in an ongoing conversation
  • Real-time coordination tracking: Live visualization of agent interactions, votes, and decision-making processes
  • Real-time feedback: Displays real-time agent and system status with enhanced coordination visualization
  • Multi-line input: Use """ or ''' to enter multi-line messages
  • Slash commands:
    • /help or /h - Show available commands
    • /status - Display current system status
    • /config - Open the configuration file
    • /clear or /reset - Clear conversation history and start fresh
    • /quit, /exit, or /q - Exit the session (or press Ctrl+C)

Watch the recorded demo:

5. πŸ“Š View Results

The system provides multiple ways to view and analyze results:

Real-time Display

  • Live Collaboration View: See agents working in parallel through a multi-region terminal display
  • Status Updates: Real-time phase transitions, voting progress, and consensus building
  • Streaming Output: Watch agents' reasoning and responses as they develop

Watch an example here:

Comprehensive Logging

All sessions are automatically logged with detailed information for debugging and analysis.

Real-time Interaction:

  • Press r during execution to view the coordination table in your terminal
  • Watch agents collaborate, vote, and reach consensus in real-time
Logging Storage Structure
.massgen/
└── massgen_logs/
    └── log_YYYYMMDD_HHMMSS/           # Timestamped log directory
        β”œβ”€β”€ agent_<id>/                 # Agent-specific coordination logs
        β”‚   └── YYYYMMDD_HHMMSS_NNNNNN/ # Timestamped coordination steps
        β”‚       β”œβ”€β”€ answer.txt          # Agent's answer at this step
        β”‚       β”œβ”€β”€ context.txt         # Context available to agent
        β”‚       └── workspace/          # Agent workspace (if filesystem tools used)
        β”œβ”€β”€ agent_outputs/              # Consolidated output files
        β”‚   β”œβ”€β”€ agent_<id>.txt          # Complete output from each agent
        β”‚   β”œβ”€β”€ final_presentation_agent_<id>.txt       # Winning agent's final answer
        β”‚   β”œβ”€β”€ final_presentation_agent_<id>_latest.txt # Symlink to latest
        β”‚   └── system_status.txt       # System status and metadata
        β”œβ”€β”€ final/                      # Final presentation phase
        β”‚   └── agent_<id>/             # Winning agent's final work
        β”‚       β”œβ”€β”€ answer.txt          # Final answer
        β”‚       └── context.txt         # Final context
        β”œβ”€β”€ coordination_events.json    # Structured coordination events
        β”œβ”€β”€ coordination_table.txt      # Human-readable coordination table
        β”œβ”€β”€ vote.json                   # Final vote tallies and consensus data
        β”œβ”€β”€ massgen.log                 # Complete debug log (or massgen_debug.log in debug mode)
        β”œβ”€β”€ snapshot_mappings.json      # Workspace snapshot metadata
        └── execution_metadata.yaml     # Query, config, and execution details
Key Log Files
  • Coordination Table (coordination_table.txt): Complete visualization of multi-agent coordination with event timeline, voting patterns, and consensus building
  • Coordination Events (coordination_events.json): Structured JSON log of all events (started_streaming, new_answer, vote, restart, final_answer)
  • Vote Summary (vote.json): Final vote tallies, winning agent, and consensus information
  • Execution Metadata (execution_metadata.yaml): Original query, timestamp, configuration, and execution context for reproducibility
  • Agent Outputs (agent_outputs/): Complete output history and final presentations from all agents
  • Debug Log (massgen.log): Complete system operations, API calls, tool usage, and error traces (use --debug for verbose logging)

β†’ For comprehensive logging guide and debugging techniques, see Logging & Debugging


πŸ€– Automation & LLM Integration

β†’ For LLM agents: See AI_USAGE.md for complete command-line usage guide

MassGen provides automation mode designed for LLM agents and programmatic workflows:

Quick Start - Automation Mode

# Run with minimal output and status tracking
uv run massgen --automation --config your_config.yaml "Your question"

Comprehensive Guide

β†’ Full automation guide with examples: Automation Guide

Topics covered:

  • Complete automation patterns with error handling
  • Parallel experiment execution
  • Performance tips and troubleshooting

Python API & LiteLLM

Use MassGen programmatically with the familiar LiteLLM/OpenAI interface:

from dotenv import load_dotenv
load_dotenv()  # Load API keys from .env

import litellm
from massgen import register_with_litellm

register_with_litellm()

# Multi-agent with slash format: "backend/model"
response = litellm.completion(
    model="massgen/build",
    messages=[{"role": "user", "content": "Compare AI approaches"}],
    optional_params={"models": ["openai/gpt-5", "groq/llama-3.3-70b"]}
)
print(response.choices[0].message.content)  # Final consensus answer

Or use the direct Python API:

from dotenv import load_dotenv
load_dotenv()

import asyncio
import massgen

result = asyncio.run(massgen.run(
    query="What is machine learning?",
    models=["openai/gpt-5", "gemini/gemini-3-pro-preview"]
))
print(result["final_answer"])  # Consensus answer from winning agent

Full API reference: Programmatic API Guide


πŸ’‘ Case Studies

To see how MassGen works in practice, check out these detailed case studies based on real session logs:

Featured:

All Case Studies:


πŸ—ΊοΈ Roadmap

MassGen is currently in its foundational stage, with a focus on parallel, asynchronous multi-agent collaboration and orchestration. Our roadmap is centered