memvid-mcp-server
A Streamable-HTTP MCP Server that uses memvid to encode text data into videos that can be quickly looked up with semantic search.
Supported Actions:
add_chunks: Adds chunks to the memory video. Note: each time you add chunks, it resets the memory.mp4. Unsure if there is a way to incrementally add.search: queries for the top-matching chunks. Returns 5 by default, but can be changed with top_k param.
Running
Set up your environment:
python3.11 -m venv my_env
. ./my_env/bin/activate
pip install -r requirements.txt
Run the server:
python server.py
With a custom port:
PORT=3002 python server.py
Connect a Client
You can connect a client to your MCP Server once it's running. Configure per the client's configuration. There is the mcp-config.json that has an example configuration that looks like this:
{
"mcpServers": {
"memvid": {
"type": "streamable-http",
"url": "http://localhost:3000"
}
}
}
Acknowledgements
- Obviously the modelcontextprotocol and Anthropic teams for the MCP Specification. https://modelcontextprotocol.io/introduction
- HeyFerrante for enabling and sponsoring this project.
Recommend MCP Servers 💡
routine-mcp-server
The Routine Model Context Protocol (MCP) server integrates with the Routine application to provide contextual information to MCP clients via stdin/stdout.
Shortcut
Shortcut is a project management platform that provides an MCP server to connect with AI tools like Cursor and Claude Code, enabling AI teammates to assist with tasks, PRs, and progress tracking.
obsidian-mcp
An MCP server that enables AI assistants to interact with Obsidian vaults, providing tools for reading, creating, editing and managing notes and tags.
code-indexer
A Model Context Protocol server that enables large language models (LLMs) to index, search, and analyze code in project directories.
bocha-search-mcp
博查是一个给AI用的搜索引擎,让你的AI应用从近百亿网页和生态内容源中获取高质量的世界知识,涵盖天气、新闻、百科、医疗、火车票、图片等多种领域。
deepseek-thinker-mcp
MCP provider that connects Deepseek reasoning content to MCP-enabled AI Clients, supporting Deepseek API and local Ollama server access.