inmemoria

Use when building persistent codebase intelligence for AI agents or integrating knowledge systems via MCP

$ Instalar

git clone https://github.com/zenobi-us/dotfiles /tmp/dotfiles && cp -r /tmp/dotfiles/ai/files/skills/projectmanagement/inmemoria ~/.claude/skills/dotfiles

// tip: Run this command in your terminal to install the skill


name: inmemoria description: Use when building persistent codebase intelligence for AI agents or integrating knowledge systems via MCP

In Memoria: Persistent Codebase Intelligence

In Memoria is an MCP server that learns your codebase patterns once, then exposes that intelligence to AI agents persistently. Instead of re-analyzing code on every interaction, it maintains a semantic understanding of your architecture, conventions, and decisions.

Core Concept

Setup → Learn → Verify → Serve. After that, AI agents query persistent intelligence without repeated parsing.

Quick Start (5 minutes)

# 1. Configure for your project
npx in-memoria setup --interactive

# 2. Build intelligence database
npx in-memoria learn ./src

# 3. Verify it worked
npx in-memoria check ./src --verbose

# 4. Keep it fresh (optional but recommended)
npx in-memoria watch ./src

# 5. Expose to agents via MCP
npx in-memoria server

When to Use

Use In Memoria:

  • Building long-lived AI agent partnerships (Claude, Copilot, etc.)
  • Projects where consistency across sessions matters
  • Teams wanting shared codebase intelligence

Skip it:

  • One-off analysis (use npx in-memoria analyze [path] directly)
  • Simple projects agents can read directly

The 5 Core Commands

CommandPurposeWhen
setup --interactiveConfigure exclusions, paths, preferencesFirst time only
learn [path]Build/rebuild intelligence databaseAfter setup, major refactors
check [path]Validate intelligence layerAfter learn, before server
watch [path]Auto-update intelligence on code changesDuring development (optional)
serverStart MCP server for agent queriesAfter check passes

Key difference: learn builds persistent knowledge. analyze is one-time reporting only.

What Agents See

When connected, agents can query:

  • Project structure - Tech stack, entry points, architecture
  • Code patterns - Your naming conventions, error handling, patterns used
  • Smart routing - "Add password reset" → suggests src/auth/password-reset.ts
  • Semantic search - Find code by meaning, not keywords
  • Work context - Track decisions, tasks, approach consistency

Troubleshooting

IssueFix
Learn failsVerify path is correct; check file permissions
Check reports missing intelligenceRun learn [path] again
Agent doesn't see new codeIs watch running? Start it: npx in-memoria watch ./src
Server won't startRun check --verbose first; if issues, rebuild: rm .in-memoria/*.db && npx in-memoria learn ./src
Multiple projects conflictUse server --port 3001 (or different port per project)

Performance Notes

  • Small projects (<1K files): 5-15s to learn
  • Medium (1K-10K files): 30-60s
  • Large (10K+ files): 2-5min

If learning stalls (>10min), verify you're not indexing node_modules/, dist/, or build artifacts—use setup's exclusion patterns.

Key Principles

  1. Local-first - Everything stays on your machine; no telemetry
  2. Persistent - One learning pass; intelligence updates incrementally with watch
  3. Agent-native - Designed for MCP; works with Claude, Copilot, and any MCP-compatible tool
  4. Pattern-based - Learns from your actual code, not rules you define

Deployment Pattern (3 terminals)

# Terminal 1: One-time setup
npx in-memoria setup --interactive
npx in-memoria learn ./src
npx in-memoria check ./src --verbose

# Terminal 2: Keep intelligence fresh
npx in-memoria watch ./src

# Terminal 3: Expose to agents
npx in-memoria server

# Now agents (Claude, Copilot, etc.) have persistent codebase context

See GitHub for full API docs and agent integration examples.