Your Second Brain, Supercharged.
MemWyre is a universal memory layer for AI—designed to capture, organize, and retrieve your knowledge across tools, conversations, and large language models.
Instead of losing context every time you switch between ChatGPT, Gemini, Claude, editors, or agents, MemWyre becomes the persistent brain that follows you everywhere.
Today’s AI tools are powerful—but stateless. Every chat starts from zero Important insights get lost in conversations Knowledge is scattered across PDFs, chats, docs, videos, and web pages Each LLM lives in its own silo Users are forced to re-explain themselves endlessly Existing tools solve storage or note-taking, but not memory continuity for AI. AI forgot. Again.
MemWyre introduces a shared, intelligent memory system that sits outside any single LLM and works with all of them.
It enables: Long-term memory for AI workflows Cross-LLM knowledge reuse Contextual retrieval grounded in your own data A clean separation between thinking (LLMs) and remembering (MemWyre) MemWyre doesn’t replace AI tools—it augments them.
Store everything in one place:
Chat outputs Documents (PDF, DOC, Markdown) Web pages & research Notes, ideas, decisions Agent outputs
All content is automatically processed, chunked, embedded, and indexed for semantic retrieval.
MemWyre goes beyond dumping files: Semantic chunking Auto-generated embeddings Metadata & source tracking Context-aware retrieval Inbox-based memory approval flow Memory becomes queryable intelligence, not dead notes.
Ask questions and get answers grounded in your own memory, not hallucinations.
Vector search Top-K relevance ranking Source-linked context Optional summaries
Works as a standalone search or as context injected into LLM prompts.
MemWyre is not tied to any single model. It integrates via: API-based LLM connectors MCP servers (for editors & agents) Browser extensions Prompt-level injection (for restricted platforms)
Your memory works across ChatGPT, Gemini, Claude, local models, and future agents.
Not all memory should be saved blindly. MemWyre introduces an Inbox model Review before committing to long-term memory Auto-approve trusted sources Manual control where it matters Prevents memory pollution
This keeps the system useful, not noisy.
Bring knowledge from everywhere:
File uploads Browser extension Web scraping YouTube transcripts Agent outputs Code editor integrations
All paths lead to the same memory engine.
- Python 3.10+
- Node.js 18+
- Chrome/Edge browser (for extension)
cd backend
python -m venv venv
venv\Scripts\activate # Windows
# source venv/bin/activate # Linux/Mac
pip install -r requirements.txt
python -m uvicorn app.main:app --reloadBackend runs at http://localhost:8000
Requires a running Redis instance.
cd backend
venv\Scripts\activate
# Windows (requires -P solo)
celery -A app.celery_app worker --loglevel=info -P solocd frontend
npm install
npm run devFrontend runs at http://localhost:5173
- Navigate to
http://localhost:5173 - Register a new account
- Login with your credentials
- Upload documents or create memories via the dashboard
- Click "Generate Prompt" on the dashboard
- Enter your query (e.g., "Summarize my project notes")
- Select a template and adjust context size
- Click "Generate Prompt"
- Copy and paste into any LLM
- Open Chrome and go to
chrome://extensions/ - Enable Developer mode
- Click Load unpacked
- Select
extensionfolder from the project root
- Click the extension icon
- Go to Settings in Brain Vault web app
- Click "Copy Extension Token"
- Paste token into extension popup
- Use the popup to generate prompts or let it auto-inject on supported sites
Add to your Claude Desktop config (~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"brain-vault": {
"command": "python",
"args": ["C:/Users/himan/OneDrive/Documents/brain_vault/backend/mcp_server.py"]
}
}
}Configure your MCP client to run:
python backend/mcp_server.pysearch_memory(query, top_k)- Search your knowledge basesave_memory(text, tags)- Save new memoriesget_document(doc_id)- Retrieve full document content
brain_vault/
├── backend/ # FastAPI server
│ ├── app/
│ │ ├── routers/ # API endpoints
│ │ ├── models/ # Database models
│ │ ├── services/ # Business logic
│ │ └── core/ # Config & security
│ ├── mcp_server.py # MCP server
│ └── requirements.txt
├── frontend/ # Vue.js app
│ ├── src/
│ │ ├── views/ # Pages
│ │ ├── components/# UI components
│ │ └── services/ # API client
│ └── package.json
└── extension/ # Chrome extension
├── manifest.json
├── background.js
├── content.js
└── popup.html
Set your LLM API keys in Settings:
- OpenAI API Key (for GPT models)
- Gemini API Key (for Gemini models)
Keys are stored locally in browser localStorage.
SQLite database is created automatically at backend/brain_vault.db.
ChromaDB stores embeddings at backend/chroma_db/.
cd backend
venv\Scripts\activate
python test_mcp_live.py- Navigate to
/promptsin the web app - Enter a test query
- Verify prompt is generated with context
Once the backend is running, visit:
- Swagger UI:
http://localhost:8000/docs - ReDoc:
http://localhost:8000/redoc
Backend:
- FastAPI
- SQLAlchemy
- ChromaDB
- LangChain
- MCP (Model Context Protocol)
Frontend:
- Vue 3
- Vite
- Tailwind CSS
- Monaco Editor
Extension:
- Manifest V3
- Chrome Extension APIs
This is a personal project, but suggestions are welcome via issues.
MIT License - feel free to use and modify.
- Built with assistance from Antigravity AI
- MCP specification by Anthropic
- Inspired by personal knowledge management best practices
