Muninn MCP Server

Local setup required. This server has to be cloned and prepared on your machine before you register it in Claude Code.
1

Set the server up locally

Run this once to clone and prepare the server before adding it to Claude Code.

Run in terminal
git clone https://github.com/wjohns989/Muninn.git
cd Muninn
pip install -e .
2

Register it in Claude Code

After the local setup is done, run this command to point Claude Code at the built server.

Run in terminal
claude mcp add -e "MUNINN_AUTH_TOKEN=${MUNINN_AUTH_TOKEN}" muninn -- python "<FULL_PATH_TO_MUNINN>/dist/index.js"

Replace <FULL_PATH_TO_MUNINN>/dist/index.js with the actual folder you prepared in step 1.

Required:MUNINN_AUTH_TOKEN+ 1 optional
README.md

Local-first persistent memory infrastructure for coding agents

Muninn

"Muninn flies each day over the world to bring Odin knowledge of what happens." โ€” Prose Edda

Local-first persistent memory infrastructure for coding agents and MCP-compatible tools.

Muninn provides deterministic, explainable memory retrieval with robust transport behavior and production-grade operational controls. Designed for long-running development workflows where continuity, auditability, and measurable quality matter โ€” across sessions, across assistants, and across projects.


๐Ÿšฉ Status

Current Version: v3.24.0 (Phase 26 COMPLETE) Stability: Production Beta Test Suite: 1422+ passing, 0 failing

What's New in v3.24.0

  • Cognitive Architecture (CoALA): Integration of a proactive reasoning loop bridging memory with active decision-making.
  • Knowledge Distillation: Background synthesis of episodic memories into structured semantic manuals for long-term wisdom.
  • Epistemic Foraging: Active inference-driven search to resolve ambiguities and fill information gaps autonomously.
  • Omission Filtering: Automated detection of missing context required for successful task execution.
  • Elo-Rated SNIPS Governance: Dynamic memory retention system mapping retrieval success to Elo ratings for usage-driven decay.

Previous Milestones

Version Phase Key Feature
v3.24.0 26 Cognitive Architecture Complete
v3.23.0 23 Elo-Rated SNIPS Governance
v3.22.0 22 Temporal Knowledge Graph
v3.19.0 20 Multimodal Hive Mind Operations
v3.18.3 19 Bulk legacy import, NLI conflict detection, uncapped discovery
v3.18.1 19 Scout synthesis, hunt mode

๐Ÿš€ Features

Core Memory Engine

  • Local-First: Zero cloud dependency โ€” all data stays on your machine
  • Multimodal: Native support for Text, Image, Audio, Video, and Sensor data
  • 5-Signal Hybrid Retrieval: Dense vector ยท BM25 lexical ยท Graph traversal ยท Temporal relevance ยท Goal relevance
  • Explainable Recall Traces: Per-signal score attribution on every search result
  • Bi-Temporal Reasoning: Support for "Valid Time" vs "Transaction Time" via Temporal Knowledge Graph
  • Project Isolation: scope="project" memories never cross repo boundaries; scope="global" memories are always available
  • Cross-Session Continuity: Memories survive session ends, assistant switches, and tool restarts
  • Bi-Temporal Records: created_at (real-world event time) vs ingested_at (system intake time)

Memory Lifecycle

  • Elo-Rated Governance: Dynamic retention driven by retrieval feedback (SNIPS) and usage statistics
  • Consolidation Daemon: Background process for decay, deduplication, promotion, and shadowing โ€” inspired by sleep consolidation
  • Zero-Trust Ingestion: Isolated subprocess parsing for PDF/DOCX to neutralize document-based exploits
  • ColBERT Multi-Vector: Native Qdrant multi-vector storage for MaxSim scoring
  • NL Temporal Query Expansion: Natural-language time phrases ("last week", "before the refactor") parsed into structured time ranges
  • Goal Compass: Retrieval signal for project objectives and constraint drift
  • NLI Conflict Detection: Transformer-based contradiction detection (cross-encoder/nli-deberta-v3-small) for memory integrity
  • Bulk Legacy Import: One-click ingestion of all discovered legacy sources (batched, error-isolated) via dashboard or API

Operational Controls

  • MCP Transport Hardening: Framed + line JSON-RPC, timeout-window guardrails, protocol negotiation
  • Runtime Profile Control: get_model_profiles / set_model_profiles for dynamic model routing
  • Profile Audit Log: Immutable event ledger for profile policy mutations
  • Browser Control Center: Web UI for search, ingestion, consolidation, and admin at http://localhost:42069
  • OpenTelemetry: GenAI semantic convention tracing (feature-gated via MUNINN_OTEL_ENABLED)

Multi-Assistant Interop

  • Handoff Bundles: Export/import memory checkpoints with checksum verification and idempotent replay
  • Legacy Migration: Discover and import memories from prior assistant sessions (JSONL chat history, SQLite state) โ€” uncapped provider limits
  • Bulk Import: POST /ingest/legacy/import-all ingests all discovered sources in batches of 50 with per-batch error isolation
  • Hive Mind Federation: Push-based low-latency memory synchronization across assistant runtimes
  • MCP 2025-11 Compliant: Full protocol negotiation, lifecycle gating, schema annotations

Quick Start

git clone https://github.com/wjohns989/Muninn.git
cd Muninn
pip install -e .

Set the auth token (shared between server and MCP wrapper):

# Windows (persists across sessions)
setx MUNINN_AUTH_TOKEN "your-token-here"

# Linux/macOS
export MUNINN_AUTH_TOKEN="your-token-here"

Start

Tools (2)

get_model_profilesRetrieves current model routing profiles.
set_model_profilesUpdates dynamic model routing profiles.

Environment Variables

MUNINN_AUTH_TOKENrequiredAuthentication token shared between server and MCP wrapper
MUNINN_OTEL_ENABLEDEnables OpenTelemetry GenAI semantic convention tracing

Configuration

claude_desktop_config.json
{"mcpServers": {"muninn": {"command": "python", "args": ["/path/to/Muninn/main.py"], "env": {"MUNINN_AUTH_TOKEN": "your-token-here"}}}}

Try it

โ†’Search my memory for the architectural decisions made during the last refactor of the authentication module.
โ†’Find all project-scoped memories related to the API integration and summarize the current constraints.
โ†’Import my legacy chat history from the previous project session into the current memory store.
โ†’What are the outstanding tasks or unresolved ambiguities identified in my recent development sessions?

Frequently Asked Questions

What are the key features of Muninn?

5-signal hybrid retrieval combining vector, lexical, graph, and temporal signals. Bi-temporal knowledge graph supporting valid time vs transaction time. Elo-rated memory governance for dynamic retention and decay. Cross-session and cross-project memory continuity. Native support for multimodal data including text, image, audio, and video.

What can I use Muninn for?

Maintaining long-term context for complex, multi-week software development projects. Synchronizing memory states across different AI assistant runtimes. Auditing and tracing AI decision-making through explainable recall traces. Consolidating fragmented project documentation into structured semantic manuals.

How do I install Muninn?

Install Muninn by running: git clone https://github.com/wjohns989/Muninn.git && cd Muninn && pip install -e .

What MCP clients work with Muninn?

Muninn works with any MCP-compatible client including Claude Desktop, Claude Code, Cursor, and other editors with MCP support.

Turn this server into reusable context

Keep Muninn docs, env vars, and workflow notes in Conare so your agent carries them across sessions.

Need the old visual installer? Open Conare IDE.
Open Conare