Skip to content

Quick Start

To use local embeddings, pull an Ollama model first (~270MB, free, no GPU required):

Terminal window
ollama pull nomic-embed-text

Then install memtomem:

Terminal window
uv tool install memtomem # or: pipx install memtomem

To also use the STM proxy:

Terminal window
uv tool install memtomem-stm # or: pip install memtomem-stm

No GPU? Pick OpenAI in the setup wizard — see the Embeddings guide.

Run the interactive 8-step wizard:

Terminal window
mm init # interactive setup
mm init -y # auto-accept defaults (for CI)
Terminal window
claude mcp add memtomem -s user -- memtomem-server

With STM proxy:

Terminal window
claude mcp add memtomem-stm -s user -- memtomem-stm

Register memtomem-server or memtomem-stm as a stdio MCP server in your editor’s MCP settings. See MCP Client Setup for details.

Use memory through your AI agent:

Tell your agentMCP tool called
”Check server status”mem_status
”Index my ~/notes folder”mem_index(path="~/notes")
”Search for deployment”mem_search(query="deployment checklist")
”Remember this insight”mem_add(content="...", tags=["ops"])
PackageDescription
memtomemCore — MCP server, CLI (mm), Web UI, hybrid search, storage
memtomem-stmSTM proxy — proactive memory surfacing via tool interception

The two packages have no Python-level dependency on each other. All communication between STM and LTM happens over the MCP protocol — they can be deployed and upgraded independently.