Commit Graph

9 Commits

Author SHA1 Message Date
Claude Code
281f05f23a Update documentation for mem0ai v1.0.0 upgrade
- Bump version to 1.2.0
- Update architecture diagram to show v1.0.0
- Update Technology Stack with new versions
- Add recent updates about v1.0.0 upgrade
- Document Neo4j 5.28+ requirement
- Document OpenAI 2.x support

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-16 21:55:46 +02:00
Claude Code
fd404d2ec5 Upgrade mem0ai to v1.0.0 with enhanced features
Major version upgrade from v0.1.118 to v1.0.0 bringing:
- Async mode as default for better performance
- Improved memory retrieval prompts
- v1.1 API format support (already configured in config.py)
- Better vector store integrations
- Enhanced LLM provider support
- Reranker support for multiple platforms

Updated dependencies:
- mem0ai[graph]: 0.1.118 → 1.0.0
- neo4j: 5.26.* → 5.28.2 (required by langchain-memgraph)
- openai: >=1.90.0,<1.110.0 → >=1.90.0 (no upper limit)

All tests passing with Ollama models (llama3.1:8b, nomic-embed-text).
Configuration already includes version: v1.1 for new API format.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-16 21:54:48 +02:00
Claude Code
eb5daa46d8 Update documentation with timezone fix and version bump
- Bump version to 1.1.1
- Update last updated date to 2025-10-16
- Add recent updates for timezone fix and ollama library

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-16 07:55:24 +02:00
Claude Code
56aa8699cc Fix timezone configuration and Ollama dependencies
- Fix mem0 library hardcoded US/Pacific timezone in Docker build
- Add TZ=Europe/Prague environment variable to containers
- Add missing ollama Python library to requirements.txt
- Add Ollama environment variables to MCP container
- Include test scripts for Ollama configuration validation

This resolves timestamp issues where memories were created with
incorrect Pacific timezone (-07:00) instead of local time (+02:00).

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-16 07:54:41 +02:00
Claude Code
fa9d3d8a6b Add Ollama support for local LLM models (Phase 2 complete)
Major Changes:
- Added Ollama as alternative LLM provider to OpenAI
- Implemented flexible provider switching via environment variables
- Support for multiple embedding models (OpenAI and Ollama)
- Created comprehensive Ollama setup guide

Configuration Changes (config.py):
- Added LLM_PROVIDER and EMBEDDER_PROVIDER settings
- Added Ollama configuration: base URL, LLM model, embedding model
- Modified get_mem0_config() to dynamically switch providers
- OpenAI API key now optional when using Ollama
- Added validation to ensure required keys based on provider

Supported Configurations:
1. Full OpenAI (default):
   - LLM_PROVIDER=openai
   - EMBEDDER_PROVIDER=openai

2. Full Ollama (local):
   - LLM_PROVIDER=ollama
   - EMBEDDER_PROVIDER=ollama

3. Hybrid configurations:
   - Ollama LLM + OpenAI embeddings
   - OpenAI LLM + Ollama embeddings

Ollama Models Supported:
- LLM: llama3.1:8b, llama3.1:70b, mistral:7b, codellama:7b, phi3:3.8b
- Embeddings: nomic-embed-text, mxbai-embed-large, all-minilm

Documentation:
- Created docs/setup/ollama.mdx - Complete Ollama setup guide
  - Installation methods (host and Docker)
  - Model selection and comparison
  - Docker Compose configuration
  - Performance tuning and GPU acceleration
  - Migration guide from OpenAI
  - Troubleshooting section
- Updated README.md with Ollama features
- Updated .env.example with provider selection
- Marked Phase 2 as complete in roadmap

Environment Variables:
- LLM_PROVIDER: Select LLM provider (openai/ollama)
- EMBEDDER_PROVIDER: Select embedding provider (openai/ollama)
- OLLAMA_BASE_URL: Ollama API endpoint (default: http://localhost:11434)
- OLLAMA_LLM_MODEL: Ollama model for text generation
- OLLAMA_EMBEDDING_MODEL: Ollama model for embeddings
- MEM0_EMBEDDING_DIMS: Must match embedding model dimensions

Breaking Changes:
- None - defaults to OpenAI for backward compatibility

Migration Notes:
- When switching from OpenAI to Ollama embeddings, existing embeddings
  must be cleared due to dimension changes (1536 → 768 for nomic-embed-text)
- Update MEM0_EMBEDDING_DIMS to match chosen embedding model

Benefits:
 Cost savings - no API costs with local models
 Privacy - all data stays local
 Offline capability - works without internet
 Model variety - access to many open-source models
 Flexibility - easy switching between providers

Version: 1.1.0
Status: Phase 2 Complete - Production Ready with Ollama Support

🤖 Generated with Claude Code

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-15 16:07:17 +02:00
Claude Code
1998bef6f4 Add MCP HTTP/SSE server and complete n8n integration
Major Changes:
- Implemented MCP HTTP/SSE transport server for n8n and web clients
- Created mcp_server/http_server.py with FastAPI for JSON-RPC 2.0 over HTTP
- Added health check endpoint (/health) for container monitoring
- Refactored mcp-server/ to mcp_server/ (Python module structure)
- Updated Dockerfile.mcp to run HTTP server with health checks

MCP Server Features:
- 7 memory tools exposed via MCP (add, search, get, update, delete)
- HTTP/SSE transport on port 8765 for n8n integration
- stdio transport for Claude Code integration
- JSON-RPC 2.0 protocol implementation
- CORS support for web clients

n8n Integration:
- Successfully tested with AI Agent workflows
- MCP Client Tool configuration documented
- Working webhook endpoint tested and verified
- System prompt optimized for automatic user_id usage

Documentation:
- Created comprehensive Mintlify documentation site
- Added docs/mcp/introduction.mdx - MCP server overview
- Added docs/mcp/installation.mdx - Installation guide
- Added docs/mcp/tools.mdx - Complete tool reference
- Added docs/examples/n8n.mdx - n8n integration guide
- Added docs/examples/claude-code.mdx - Claude Code setup
- Updated README.md with MCP HTTP server info
- Updated roadmap to mark Phase 1 as complete

Bug Fixes:
- Fixed synchronized delete operations across Supabase and Neo4j
- Updated memory_service.py with proper error handling
- Fixed Neo4j connection issues in delete operations

Configuration:
- Added MCP_HOST and MCP_PORT environment variables
- Updated .env.example with MCP server configuration
- Updated docker-compose.yml with MCP container health checks

Testing:
- Added test scripts for MCP HTTP endpoint verification
- Created test workflows in n8n
- Verified all 7 memory tools working correctly
- Tested synchronized operations across both stores

Version: 1.0.0
Status: Phase 1 Complete - Production Ready

🤖 Generated with Claude Code

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-15 13:56:41 +02:00
Claude Code
9bca2f4f47 Add implementation summary document 2025-10-14 08:45:22 +02:00
Claude Code
61a4050a8e Complete implementation: REST API, MCP server, and documentation
Implementation Summary:
- REST API with FastAPI (complete CRUD operations)
- MCP Server with Python MCP SDK (7 tools)
- Supabase migrations (pgvector setup)
- Docker Compose orchestration
- Mintlify documentation site
- Environment configuration
- Shared config module

REST API Features:
- POST /v1/memories/ - Add memory
- GET /v1/memories/search - Semantic search
- GET /v1/memories/{id} - Get memory
- GET /v1/memories/user/{user_id} - User memories
- PATCH /v1/memories/{id} - Update memory
- DELETE /v1/memories/{id} - Delete memory
- GET /v1/health - Health check
- GET /v1/stats - Statistics
- Bearer token authentication
- OpenAPI documentation

MCP Server Tools:
- add_memory - Add from messages
- search_memories - Semantic search
- get_memory - Retrieve by ID
- get_all_memories - List all
- update_memory - Update content
- delete_memory - Delete by ID
- delete_all_memories - Bulk delete

Infrastructure:
- Neo4j 5.26 with APOC/GDS
- Supabase pgvector integration
- Docker network: localai
- Health checks and monitoring
- Structured logging

Documentation:
- Introduction page
- Quickstart guide
- Architecture deep dive
- Mintlify configuration

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-14 08:44:16 +02:00
Claude Code
cfa7abd23d Initial commit: Project foundation and architecture
- Add project requirements document
- Add comprehensive architecture design
- Add README with quick start guide
- Add .gitignore for Python/Docker/Node

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2025-10-13 15:01:50 +02:00