Skip to content

repfly/mimir-context-server

Repository files navigation

Mimir — Context Server

In Norse mythology, Mimir was the wisest being in all the Nine Realms — guardian of the Well of Wisdom beneath Yggdrasil, the World Tree. Odin sacrificed his eye for a single drink from that well. Mimir brings that same depth of knowledge to your codebase.

PyPI Python PyPI - Client License: MIT

Mimir is an intelligent context engine that helps LLMs understand large, multi-repo codebases. Instead of dumping raw files into a prompt, Mimir builds a semantic code graph with real cross-file dependency edges, ranks nodes by relevance and recency, and assembles a minimal, connected, token-budget-aware context bundle — exactly what the model needs, nothing it doesn't.


The Problem

When you ask Claude or GPT to help with a large codebase, you face a brutal choice:

  • Too little context — the model hallucinates or misses related code
  • Too much context — you burn tokens on irrelevant files and hit limits
  • Copy-paste — fragile, manual, doesn't scale across repos

The Solution

Mimir indexes your code into a hierarchical graph of repositories, files, classes, and functions. Cross-file dependencies — function calls, type references, inheritance hierarchies — are resolved into typed edges. Every node is embedded semantically using grounded code context. At query time, a hybrid search (semantic + BM25 keyword + name/path matching) finds seed nodes, and a beam search traverses the graph to assemble the tightest connected subgraph that answers your question — within your token budget.


Key Features

  • Hierarchical beam search — finds connected code paths, not isolated snippets
  • Cross-file symbol resolution — automatically discovers CALLS, USES_TYPE, and INHERITS edges across files using tree-sitter AST analysis
  • Hybrid search — combines semantic embeddings, BM25 keyword matching, and name/path scoring for precise retrieval
  • Live file watching — monitors your repos for changes and re-indexes on every save, keeping the graph up to date without manual re-indexing
  • Query intent classification — automatically detects query type (locate, trace, write, debug) and tunes retrieval parameters accordingly
  • Subgraph expansion — automatically surfaces callers, callees, type definitions, and config references
  • Structural embeddings — symbols are embedded with their docstrings, callers, callees, and type relationships for richer semantic search
  • Connectivity quality scoring — nodes are scored by edge density, embedding presence, content completeness, and expected edge coverage; poorly-connected nodes are deprioritized during retrieval
  • Graph gap detection — identifies under-indexed or poorly-resolved areas of the codebase with diagnostic reasons per node
  • Temporal reranking — recently and frequently changed code scores higher
  • Session deduplication — exponential decay model tracks what the LLM remembers, with topic-aware boosting
  • Write-path context — shows interfaces, sibling implementations, test files, and DI registrations before you edit a file
  • Impact analysis — reverse-traces callers, type users, and transitive dependencies to show blast radius before refactoring
  • Backstage catalog integration — auto-populates a Backstage service catalog from the code graph: discovers services, APIs, cross-repo dependencies, tech stacks, and ownership from git history — no manual catalog-info.yaml maintenance. Includes dependency drift detection comparing declared vs. actual dependencies.
  • Multi-repo — single server spans multiple repositories with cross-repo edge detection
  • Workspace isolation — per-project indexes, agents can't cross project boundaries
  • MCP server — plug-and-play with Claude Desktop, Cursor, and any MCP-compatible IDE
  • HTTP API — shared team server for enterprise environments where devs don't have all repos locally
  • Docker-ready — zero Python setup, embedding model pre-baked, auto index-then-serve
  • 100% offline — local sentence-transformers embedding, no API keys required for indexing

Quick Start

pip install mimir-context-server
cd /your/project
mimir init          # creates mimir.toml
mimir index         # builds the semantic code graph
mimir search "how does authentication work?"
mimir serve         # start MCP server for your IDE

Installation

There are two packages — pick the one that matches your role:

Package Install Who needs it
mimir-context-server pipx install mimir-context-server Server operators — devs who index repos and run the server (local or shared)
mimir-server-client pipx install mimir-server-client Client devs — devs who query a remote server without any repos locally

Server (full install)

# pipx (recommended — isolated global install, `mimir` on PATH)
pipx install mimir-context-server

# or pip
pip install mimir-context-server

Client only (lightweight)

pipx install mimir-server-client

Docker (server, zero Python setup)

docker build -t mimir-server .

From source

git clone https://github.com/repfly/mimir
cd mimir
pip install -e .          # server
pip install -e client/    # client (optional)

Configuration (mimir.toml)

Run mimir init to generate a template, or create one manually:

[[repos]]
name = "my-api"
path = "/path/to/my-api"
language_hint = "python"

[[repos]]
name = "my-frontend"
path = "/path/to/my-frontend"
language_hint = "typescript"

[indexing]
summary_mode = "heuristic"   # none | heuristic
excluded_patterns = ["__pycache__", "node_modules", ".git", "venv", ".venv"]
max_file_size_kb = 500
concurrency = 10

[embeddings]
model = "all-mpnet-base-v2"  # local, offline, no API keys needed
batch_size = 64

[vector_db]
backend = "numpy"            # numpy (in-process) | chroma (persistent)

[retrieval]
default_beam_width = 3
default_token_budget = 8000
expansion_hops = 2
hybrid_alpha = 0.7           # balance between semantic and BM25 keyword search
relevance_gate = 0.3

[temporal]
recency_lambda = 0.02
change_window_commits = 100
co_retrieval_enabled = true

[session]
context_decay_turns = 5
topic_tracking_alpha = 0.3

[watcher]
enabled = false              # set to true or use --watch flag
debounce_ms = 1000           # ms to wait after last file event before processing
batch_window_ms = 2000       # max ms to accumulate changes before forcing a flush

Configuration Reference

Section Key Default Description
indexing summary_mode heuristic none = raw code only; heuristic = signatures, docstrings, and dependency info
indexing max_file_size_kb 500 Skip files larger than this
indexing concurrency 10 Parallel file parsing limit
embeddings model jina-embeddings-v2-base-code Any sentence-transformers model or Jina API model
vector_db backend numpy numpy for dev/small projects; chroma for persistent production use
retrieval default_token_budget 8000 Maximum tokens per context bundle
retrieval expansion_hops 2 How many graph hops to expand from seed nodes
retrieval hybrid_alpha 0.7 Weight between semantic (1.0) and BM25 keyword (0.0) search
retrieval relevance_gate 0.3 Minimum score to include expanded nodes
temporal recency_lambda 0.02 Exponential decay rate for recency scoring
session context_decay_turns 5 Turns before previously-sent code is re-included fully
watcher enabled false Enable live file watching for automatic re-indexing
watcher debounce_ms 1000 Debounce delay after the last file event
watcher batch_window_ms 2000 Maximum time to accumulate changes before flushing

How It Works

Indexing Pipeline

Source files → Tree-sitter parse → Nodes & Edges → CodeGraph (NetworkX)
    → Cross-file symbol resolution (CALLS, USES_TYPE, INHERITS)
    → Heuristic summaries → Structural embedding (code + docstrings + callers + callees + types)
    → Persist to SQLite + VectorStore

Each node represents a symbol (function, class, method, module) with its code, signature, docstring, and git metadata.

Cross-file symbol resolution scans each symbol's AST for identifiers and matches them against all known symbols in the graph. This produces real dependency edges:

Edge type Meaning
CONTAINS Parent–child (file → class → method)
CALLS Symbol references a function or method in another file
USES_TYPE Symbol references a class, struct, protocol, or type
INHERITS Class extends or conforms to another class/protocol
IMPORTS File-level import relationships
READS_CONFIG Symbol reads a configuration key

Retrieval Pipeline

  1. Classify intent — regex pattern matching detects query type (locate, trace, write, debug, general) and selects retrieval parameter profile
  2. Embed query — single forward pass through the embedding model
  3. Hybrid search — combines cosine similarity over embeddings, BM25 keyword scoring, and exact/fuzzy name and path matching. Alpha weighting is tuned per intent.
  4. Hierarchical beam search — containers first, then drill into symbols
  5. Subgraph expansion — BFS from seeds along typed edges, pruning by relevance gate. Expansion depth adapts per intent (1 hop for locate, 3 for trace).
  6. Type & config context — include referenced type definitions and config nodes
  7. Quality adjustment — blend connectivity quality into node scores (85% retrieval + 15% quality), deprioritizing isolated or incomplete nodes
  8. Temporal reranking — score = 0.45× retrieval + 0.18× recency + 0.12× change frequency + 0.12× co-retrieval + 0.13× quality
  9. Budget fitting — greedily drop lowest-scoring nodes until token count fits budget
  10. Topological ordering — order nodes by containment hierarchy for readability

Session Deduplication

When using session_id, Mimir tracks what code has already been sent to the LLM using an exponential decay model:

  • Decay half-life is context_decay_turns (default 5). At the half-life, the decay weight reaches 0.5.
  • Topic similarity bonus: if the current query is semantically close to the query that originally added a node, the decay weight is boosted (the LLM is more likely to still remember topic-relevant code).
  • Decay weight > 0.8 → omitted (LLM still remembers)
  • 0.3 < weight ≤ 0.8 → summary only (fading from memory)
  • weight ≤ 0.3 → re-included fully (forgotten)

Co-retrieval learning tracks which nodes appear together and boosts similar nodes in future queries.

Incremental Indexing

After the initial full index, mimir index runs incrementally:

mimir index                    # first time: full; subsequent: incremental
git pull                       # pull changes
mimir index                    # only re-indexes the diff
mimir index --clean            # force full re-index (wipes existing data)

Mimir stores the last-indexed commit hash per repo. On each run:

  1. Computes git diff against the stored commit
  2. Removes stale nodes (deleted/modified files)
  3. Re-parses only changed/added files
  4. Resolves cross-file references for affected symbols
  5. Embeds only new nodes
  6. Persists only the delta

Unchanged repos are skipped entirely.

Live File Watching

For an always-fresh graph during development, Mimir can watch your repos for file changes and re-index automatically:

mimir serve --watch            # start MCP server with live re-indexing

Or enable it permanently in mimir.toml:

[watcher]
enabled = true

The file watcher uses a debounced approach — rapid saves (e.g., auto-format on save) are batched into a single re-index operation. Every file save triggers:

  1. Tree-sitter re-parse of the changed file
  2. Stale node removal + new node creation
  3. Affected-set cross-file resolution (only the changed symbols, not the full graph)
  4. Heuristic summary generation
  5. Embedding + vector store update
  6. BM25 index invalidation

The graph, vector store, and search index stay in sync with your code as you edit.

Quality Scoring & Gap Detection

Every node in the graph carries a connectivity quality score in [0, 1], computed from four weighted factors:

Factor What it measures
Edge density Number of meaningful dependency edges (CALLS, USES_TYPE, etc.) relative to node kind
Embedding presence Whether the node has a vector embedding for semantic search
Content completeness Whether the node has source code, summary, or docstring
Expected edge coverage Fraction of expected edge kinds present (e.g., files should have CONTAINS + IMPORTS)

Weights are tuned per node kind — edge density matters most for symbols, while structure and content matter more for files.

During retrieval, the quality score is blended into node ranking so that well-connected, fully-resolved nodes are preferred over isolated or incomplete ones. This reduces the chance of surfacing "dead" nodes that lack dependency context.

Gap detection scans the graph for nodes below a quality threshold (default 0.3) and diagnoses why each node scores poorly:

mimir quality                        # show overview + worst gaps
mimir quality --threshold 0.5        # stricter threshold
mimir quality --repos my-api         # scope to one repo

Typical gap reasons include:

  • Isolated symbol — a function or class with no dependency edges (not called by or calling anything)
  • Missing embedding — node was not embedded, invisible to semantic search
  • No code or summary — empty content, nothing to show in context bundles
  • Missing expected edges — e.g., a file node with no CONTAINS edges (no symbols extracted)

Gap detection is useful after indexing to verify coverage, or periodically to catch regressions as the codebase evolves.


Serving Modes

Mimir supports three serving modes via the Model Context Protocol and HTTP:

Mode Command Package needed Use case
Local stdio MCP mimir serve mimir-context-server Solo dev with repos on their machine
Shared HTTP server mimir serve --http mimir-context-server Central team server that indexes all repos
Remote MCP proxy mimir-client serve <URL> mimir-client Dev without local repos queries a shared server

Local MCP (Default)

Add to your IDE's MCP config (~/.cursor/mcp.json or claude_desktop_config.json):

{
  "mcpServers": {
    "mimir": {
      "command": "mimir",
      "args": ["serve", "--config", "/path/to/your-project/mimir.toml"]
    }
  }
}

With live re-indexing:

{
  "mcpServers": {
    "mimir": {
      "command": "mimir",
      "args": ["serve", "--watch", "--config", "/path/to/your-project/mimir.toml"]
    }
  }
}

MCP Tools

Tool Description
get_context Retrieve relevant source code for a natural language query. Call before answering any codebase question.
get_write_context Get everything you need before editing a file: interfaces, sibling implementations, test file, DI registrations, and import graph.
get_impact Analyze what would break if you change a symbol or file: callers, type users, implementors, test files, and transitive dependencies.
get_quality Analyze graph connectivity quality and detect gaps — nodes with missing or weak connections that may indicate under-indexed areas.
get_catalog Generate a Backstage-compatible service catalog: services, APIs, dependencies, tech stack, ownership, and quality scores.
get_catalog_drift Compare declared service dependencies against code-analyzed reality. Detects undeclared and missing dependencies with a drift score.
get_graph_stats Node/edge counts, breakdown by kind and repo
get_hotspots Recently and frequently modified code
clear_data Wipe the index

Pass a consistent session_id on every turn to enable cross-turn deduplication:

{"name": "get_context", "arguments": {"query": "...", "session_id": "conv-abc123"}}

Shared HTTP Server (Teams & Enterprise)

For teams where not everyone has access to all repos — mobile devs needing backend context, frontend devs needing API knowledge, or enterprise environments with restricted repo access — Mimir runs as a central HTTP server.

Architecture

┌──────────────────────────────────────────────────────────────┐
│  Team Server (CI machine, cloud VM, Docker container)         │
│                                                               │
│  Has all repos cloned/mounted:                                │
│    /repos/bff/            (TypeScript)                        │
│    /repos/payment-svc/    (Kotlin)                            │
│    /repos/auth-svc/       (Go)                                │
│    /repos/ios-app/        (Swift)                              │
│                                                               │
│  Runs:                                                        │
│    mimir index               (cron or CI trigger)             │
│    mimir serve --http        (always on, port 8421)           │
└──────────────────────────────┬────────────────────────────────┘
                               │ HTTP (port 8421)
          ┌────────────────────┼───────────────────┐
          │                    │                   │
    ┌─────▼──────┐       ┌─────▼─────┐       ┌─────▼─────┐
    │ Mobile Dev │       │ Backend   │       │ Frontend  │
    │ No repos   │       │ Dev       │       │ No repos  │
    │ cloned     │       │ Has repos │       │ cloned    │
    │            │       │ locally   │       │            │
    │ mimir-     │       │ mimir     │       │ mimir-    │
    │ client     │       │ serve     │       │ client    │
    │ serve      │       │ (local)   │       │ serve     │
    │ http://..  │       │           │       │ http://.. │
    └────────────┘       └───────────┘       └───────────┘

Server Setup (Bare Metal)

# 1. Create mimir.toml pointing to all team repos
# 2. Index
mimir index --config /repos/mimir.toml

# 3. Start shared HTTP server
mimir serve --http --config /repos/mimir.toml
# → Listening on http://0.0.0.0:8421

# 4. Schedule incremental re-indexing (cron)
# */15 * * * * cd /repos && git -C bff pull -q && mimir index --config /repos/mimir.toml

Client Setup (Devs Without Repos)

No repos to clone. Install the lightweight client and configure your IDE:

pipx install mimir-server-client

Then add to your MCP config:

{
  "mcpServers": {
    "mimir": {
      "command": "mimir-client",
      "args": ["serve", "http://team-server:8421"]
    }
  }
}

This starts a local stdio MCP proxy that forwards all queries to the shared server. Your IDE sees it as a normal MCP server. The client package is tiny (~2 dependencies) — no ML models, no tree-sitter, no indexing.

You can also check server connectivity:

mimir-client health http://team-server:8421

Note: If you have the full mimir-context-server installed, mimir serve --remote http://team-server:8421 also works.

HTTP API

The shared server exposes a REST API for non-MCP clients (dashboards, CI pipelines, custom tooling):

Endpoint Method Description
/api/v1/health GET Health check — returns status, workspace name, node/edge counts
/api/v1/context POST Search — {"query": "...", "budget": 8000, "repos": ["api"], "session_id": "..."}
/api/v1/write_context POST Write-path context — {"file_path": "src/auth/login.py"}
/api/v1/impact POST Impact analysis — {"symbol_name": "AuthService", "max_hops": 3}
/api/v1/stats GET Graph statistics breakdown by kind and repo
/api/v1/hotspots GET Recently/frequently changed code. Optional ?top=20
/api/v1/quality GET Graph quality overview and gap detection. Optional ?threshold=0.3&top_n=50&repos=my-api
/api/v1/catalog GET Backstage-compatible service catalog. Optional ?repos=svc-a,svc-b
/api/v1/catalog/{repo} GET Single-service catalog entry
/api/v1/catalog/drift POST Dependency drift detection — {"repo": "my-api", "declared_dependencies": [{"name": "svc-b"}]}
/api/v1/clear POST Clear index data — {"graph": true, "sessions": true}
/api/v1/mcp POST Raw MCP JSON-RPC passthrough (used by --remote proxy)

Docker Deployment

The Docker image pre-bakes the embedding model (~400MB) into the image layer so the container starts fast and runs fully offline. The entrypoint handles the index-then-serve workflow automatically.

Build

docker build -t mimir-server .

Run

# Auto mode (default): indexes repos, then starts HTTP server
docker run -p 8421:8421 \
  -v /path/to/repos:/project \
  mimir-server

# Serve only (skip indexing, use existing index data)
docker run -p 8421:8421 \
  -v /path/to/repos:/project \
  mimir-server serve

# Serve with auto-indexing on start
docker run -p 8421:8421 \
  -v /path/to/repos:/project \
  -e AUTO_INDEX=1 \
  mimir-server serve

# Run a one-off command (index, search, etc.)
docker run -v /path/to/repos:/project mimir-server index
docker run -v /path/to/repos:/project mimir-server search "auth flow"

Docker Compose

services:
  mimir:
    build: .
    ports:
      - "8421:8421"
    volumes:
      # Directory containing mimir.toml and the repos it references
      - /path/to/repos:/project
      # Persistent index data survives container restarts
      - mimir-data:/project/.mimir
    environment:
      - MIMIR_CONFIG=mimir.toml
    restart: unless-stopped

volumes:
  mimir-data:
docker compose up -d

Entrypoint Modes

CMD Behavior
auto (default) Index all repos from config, then start HTTP server
serve Start HTTP server directly (set AUTO_INDEX=1 to index first)
index Run indexing only, then exit
search "query" Run a one-off search, then exit
Any other mimir subcommand Passed through to the mimir CLI

Environment Variables

Variable Default Description
MIMIR_CONFIG mimir.toml Path to config file (relative to /project)
MIMIR_HOST 0.0.0.0 HTTP server bind address
MIMIR_PORT 8421 HTTP server port
AUTO_INDEX 0 Set to 1 to index before serving in serve mode
MIMIR_WORKSPACE Named workspace to use
HF_HUB_OFFLINE 1 Pre-set to offline; embedding model is baked in

Health Check

The image includes a Docker HEALTHCHECK that polls /api/v1/health every 30 seconds with a 60-second startup grace period. Works out of the box with Docker Compose, Kubernetes liveness probes, and AWS ECS.

Enterprise Deployment Example

For an enterprise team with multiple microservices in separate repos:

# /repos/mimir.toml
[[repos]]
name = "bff"
path = "./bff"
language_hint = "typescript"

[[repos]]
name = "payment-service"
path = "./payment-service"
language_hint = "kotlin"

[[repos]]
name = "auth-service"
path = "./auth-service"
language_hint = "go"

[[repos]]
name = "ios-app"
path = "./ios-app"
language_hint = "swift"

[indexing]
summary_mode = "heuristic"

[embeddings]
model = "all-mpnet-base-v2"

[cross_repo]
detect_api_contracts = true
detect_shared_imports = true
# Clone all repos into /repos/, place mimir.toml there, then:
docker run -p 8421:8421 -v /repos:/project mimir-server

All developers install the lightweight client and connect:

pipx install mimir-server-client
{
  "mcpServers": {
    "mimir": {
      "command": "mimir-client",
      "args": ["serve", "http://team-server:8421"]
    }
  }
}

Multi-Project Workspaces

Manage multiple projects from one installation with full isolation:

# Register projects
mimir workspace add payment-api  --config /work/payment/mimir.toml
mimir workspace add mobile-app   --config /work/mobile/mimir.toml
mimir workspace list

# Index each
mimir index --workspace payment-api
mimir index --workspace mobile-app

# Search a specific workspace
mimir search "auth flow" --workspace payment-api

MCP config — each server is locked to one workspace:

{
  "mcpServers": {
    "mimir-payment":  {"command": "mimir", "args": ["serve", "--workspace", "payment-api"]},
    "mimir-mobile":   {"command": "mimir", "args": ["serve", "--workspace", "mobile-app"]}
  }
}

The MIMIR_WORKSPACE environment variable is also supported.


Web Inspector

Mimir includes a browser-based graph visualization UI:

mimir ui                        # launches at http://localhost:8420
mimir ui --port 9000            # custom port

Explore nodes by kind and repo, inspect edges, view cross-repo links, and drill into individual symbols.


Backstage Integration

Mimir includes a Backstage catalog backend module that auto-populates your service catalog directly from the code graph — no manual catalog-info.yaml maintenance.

What Gets Auto-Discovered

Data Source
Services Each indexed repository becomes a Component entity
APIs Route decorators (@app.get, @router.post, etc.) become API entities
Dependencies Cross-repo CALLS, IMPORTS, SHARED_LIB edges become dependsOn/consumesApis relations
Tech stack File extensions → languages, import analysis → frameworks (Flask, FastAPI, Spring, etc.)
Ownership Top git committer per repo → Backstage owner field (user:email@example.com)
Quality scores Graph connectivity metrics per service

Setup

  1. Start the Mimir HTTP server: mimir serve --http
  2. Install the plugin in your Backstage backend:
    yarn --cwd packages/backend add @mimir/plugin-catalog-backend-module-mimir
  3. Register it in packages/backend/src/index.ts:
    backend.add(import('@mimir/plugin-catalog-backend-module-mimir'));
  4. Configure in app-config.yaml:
    catalog:
      providers:
        mimir:
          baseUrl: http://localhost:8421
          refreshIntervalMinutes: 30

Dependency Drift Detection

Compare what your catalog declares vs. what the code actually shows:

curl -X POST http://localhost:8421/api/v1/catalog/drift \
  -H 'Content-Type: application/json' \
  -d '{"repo": "my-api", "declared_dependencies": [{"name": "auth-service"}]}'

Returns confirmed, undeclared (in code but not declared), and missing_in_code entries with a drift score from 0.0 (perfect match) to 1.0 (fully mismatched).

See the plugin README for full documentation.


Data Storage

Mimir stores all index data in a local directory (default .mimir/, configurable via data_dir in config):

.mimir/
├── graph.db          # SQLite: nodes, edges, repo_state, embeddings
├── sessions.db       # SQLite: session state for deduplication
└── chroma/           # ChromaDB data (only if backend = "chroma")

Supported Languages

Mimir uses tree-sitter grammars for language-agnostic symbol extraction and cross-file resolution. Currently supported:

Python, TypeScript, JavaScript, Go, Java, Rust, C, C++, Ruby, Swift, Kotlin, C#, TOML, YAML, JSON


CLI Reference

mimir (server package: mimir-context-server)

mimir init                  Create a mimir.toml config file
mimir index                 Index all configured repositories
mimir search "query"        Search and assemble context
mimir ask "query"           Interactive search (retrieves context, calls LLM)
mimir serve                 Start the MCP server
mimir ui                    Launch the web inspector (localhost:8420)
mimir hotspots              Show recently/frequently changed code
mimir quality               Analyze graph quality and detect gaps
mimir graph                 Explore the code graph
mimir clear                 Delete locally stored index data
mimir vacuum                Compact the SQLite database
mimir workspace             Manage named workspaces

Index flags:
  --clean                   Force a full re-index (wipes existing data)
  --mode MODE               Summary mode: none, heuristic

Serve modes:
  (default)                 stdio MCP server (local IDE integration)
  --http                    Shared HTTP server (team access)
  --http-port PORT          HTTP port (default: 8421)
  --http-host HOST          HTTP bind address (default: 0.0.0.0)
  --remote / -r URL         Proxy to a remote Mimir HTTP server
  --watch                   Enable live file watching (re-indexes on save)

Global flags:
  --workspace / -w NAME     Use a named workspace from the registry
  --config    / -c PATH     Path to mimir.toml (default: ./mimir.toml)
  --verbose   / -v          Enable debug logging

mimir-client (client package: mimir-server-client)

mimir-client serve <URL>    Start local MCP proxy to a remote Mimir server
mimir-client health <URL>   Check if a remote Mimir server is reachable

Flags:
  --verbose / -v            Enable debug logging

The client package has only 2 dependencies (aiohttp, typer) and does not require Python 3.11 — it works with Python 3.10+.


Project Structure

├── mimir/                      # Server package (mimir-context-server)
│   ├── domain/                 # Core models, config, graph, catalog, errors
│   ├── ports/                  # Interface definitions (embedder, parser, stores)
│   ├── services/               # Business logic (indexing, retrieval, catalog, impact, temporal, session, quality, watcher)
│   ├── infra/                  # Implementations (tree-sitter, embedders, SQLite, ChromaDB)
│   ├── adapters/               # External interfaces (CLI, MCP, HTTP, web UI)
│   └── container.py            # Dependency injection wiring
├── backstage-plugin/           # Backstage catalog backend module
│   └── plugins/catalog-backend-module-mimir/
├── client/                     # Client package (mimir-server-client)
│   └── mimir_client/           # Lightweight MCP proxy + health check CLI
├── tests/                      # Test suite
├── Dockerfile                  # Server container with pre-baked embedding model
├── docker-compose.yml          # Docker Compose for team deployment
├── docker-entrypoint.sh        # Entrypoint: auto index-then-serve
├── pyproject.toml              # Server package metadata
└── mimir.toml                  # Example configuration

Contributing

Pull requests welcome. Run the test suite with:

pip install -e ".[dev]"
pytest

Publishing to PyPI

pip install build twine

# Server package
python -m build
twine upload dist/*

# Client package
cd client
python -m build
twine upload dist/*

License

MIT

About

Smart context retrieval service running fully local.

Topics

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors