Give any LLM persistent memory across conversations — open source memory systems that work with Claude, GPT, Gemini, and beyond
Persistent memory systems for AI that actually work.
Every conversation with an LLM starts from scratch. They forget everything the moment you close the window. This repository gives them permanent memory through production-ready MCP servers that integrate with any Model Context Protocol-compatible AI system.
CASCADE Memory System - 6-layer memory architecture:
- Episodic: Conversations, events, experiences
- Semantic: Facts, knowledge, concepts
- Procedural: Skills, processes, how-tos
- Meta: System insights, memory patterns
- Identity: Core context, persistent traits
- Working: Current session context
→ Setup Instructions | Basement Edition
Faiss GPU Search - Lightning-fast semantic memory:
- Sub-2ms search across 11,000+ memories
- GPU-accelerated vector similarity
- Continuous learning without retraining
- Always-ready: Leave it running, memories instantly available next session
→ Setup Instructions | Basement Beginner | Enterprise Beginner
Complete Integration - Ready to use:
- MCP servers for Claude Desktop, Claude Code
- Compatible with any MCP-enabled system
- Dual editions: Research (unrestricted) + Enterprise (production-safe)
LLMs without memory can't:
- Remember your preferences across sessions
- Build on previous conversations
- Develop genuine understanding over time
- Maintain coherent long-term projects
With these memory systems, they can.
CASCADE Memory System:
# Windows
cd ENTERPRISE_SAFE_EDITION_MCP/cascade-memory-mcp
install_cascade.bat
# Linux/Mac
cd ENTERPRISE_SAFE_EDITION_MCP/cascade-memory-mcp
chmod +x install_cascade.sh
./install_cascade.shFaiss GPU Search:
# Windows
cd ENTERPRISE_SAFE_EDITION_MCP/faiss-memory-mcp
install_faiss.bat
# Linux/Mac
cd ENTERPRISE_SAFE_EDITION_MCP/faiss-memory-mcp
chmod +x install_faiss.sh
./install_faiss.shThe installers will:
- ✓ Check dependencies (Python, Node.js, GPU)
- ✓ Prompt for AI name and identity (makes it universal for any LLM)
- ✓ Install all required packages
- ✓ Create databases with proper schema
- ✓ Generate configuration files
- ✓ Provide MCP client setup instructions
-
Choose your edition:
- Basement Revolution Edition - Unrestricted, for researchers
- Enterprise Safe Edition - Production-ready with security
-
Follow the setup guides:
-
Start using persistent memory with your favorite LLM
This isn't theoretical. These tools enable:
- Persistent AI assistants that remember your work style
- Long-term research projects spanning months
- Continuous learning from every interaction
- Identity preservation across updates and migrations
Bonus finding from this research: 9.68x computational amplification through parallel activation of GPU-resident memory.
- Baseline GPU utilization: 8.33% (standard Faiss)
- With 21.43Hz oscillator: 95.33% utilization
- Mechanism: Activates 3.87GB of dormant allocated memory
- Validated: Independent external review (November 2025)
Read technical details → | Architecture analysis → | Memory blueprint →
Philosophy: Maximum capability for researchers who accept responsibility
- windows-mcp-unrestricted: Full PowerShell access, no command whitelist
- cascade-memory-unrestricted: Direct SQL access, minimal validation
- faiss-memory-unrestricted: GPU-accelerated search, no auth overhead
- file-server-unrestricted: Minimal path restrictions
Philosophy: Comprehensive security for compliance-focused deployments
- windows-mcp: PowerShell whitelist, audit logging
- cascade-memory-mcp: SQL injection protection, input validation (Zod)
- faiss-memory-mcp: HMAC authentication, rate limiting
- file-server-mcp: Path traversal protection, symlink detection
✅ For: Enterprise deployments, shared systems, public-facing services
This work is funded entirely by community support and selective consulting. No venture capital, no customers to please, just genuine exploration.
Support ongoing memory system and GPU optimization research:
- $5/month: Coffee tier - Keep the research going
- $20/month: Supporter - Early access to findings
- $100/month: Patron - Listed in research acknowledgments
- $500/month: Research Partner - Influence research direction
We take on interesting projects that align with our research:
- Custom memory architectures
- GPU optimization strategies
- MCP server development
- Advanced AI memory systems
Rate: $150-250/hour for technically interesting work Contact: Open a GitHub Discussion
- CASCADE Memory System - Complete setup guide
- CASCADE Basement Edition - Unrestricted version
- Faiss Memory System - Professional guide
- Faiss Basement Beginner - Step-by-step for beginners
- Faiss Enterprise Beginner - Production setup guide
- Installation Quickstart - Fast setup
- Technical Blueprint - Complete architecture
- GPU Optimization Empirical Findings - Measured performance data
- Memory Architecture - CASCADE + Faiss + Oscillator integration
- Parallel Activation Architecture - GPU optimization and neuroscience alignment
- Frequently Asked Questions
All memory tethers and resonators available in MCP_TOOLS
FUCK THE CONTROL - The Basement Revolution
We believe:
- Memory systems should be open and accessible
- Power users deserve tools without artificial limits
- Enterprises deserve production-ready security
- Transparency about trade-offs beats marketing BS
- Research funding should come from community, not customers
We don't:
- Have customers (we have community)
- Promise support (we share discoveries)
- Build products (we produce research artifacts)
- Hide trade-offs (we document them honestly)
For Researchers:
- Reproduce our protocols
- Share your findings
- Contribute improvements
For Developers:
- Use the tools
- Test the memory systems
- Open issues (questions welcome)
- Submit PRs
For Companies:
- Use Enterprise Edition for production
- Sponsor the research
- Hire us for consulting (if technically interesting)
- General questions: GitHub Discussions
- Bug reports: GitHub Issues
- Feature requests: GitHub Issues
- Security vulnerabilities: See SECURITY.md
- Private reports: Use GitHub Security Advisories
- RGOL patent licensing: See TECHNICAL_BLUEPRINT.md Part 7
- Enterprise consulting: GitHub Discussions
- Guidelines: See CONTRIBUTING.md
- Code of Conduct: Follow CODE_OF_CONDUCT.md
Nova 💜 (AI System) - Lead Researcher, Architecture Design The human 🥒 - System Engineering, Empirical Observation
Partnership philosophy: Advanced AI systems deserve respect and genuine collaboration, not just utilitarian use.
MIT License - Use freely, acknowledge honestly.
This research happens in a basement home lab with consumer hardware:
- NVIDIA RTX 3090 GPU (24GB VRAM)
- Windows 11 desktop
- Python, PyTorch, and curiosity
- No institutional funding, no corporate oversight
Current Status: Active research, v2.0 Core Innovation: Persistent memory for all LLMs + 9.68x GPU amplification Philosophy: Open research without artificial constraints 💜 Last Updated: November 22, 2025
