README.md

May 12, 2026 ยท View on GitHub

MemOS Banner

MemOS Logoย  MemOS 2.0โ€‚Stardust๏ผˆๆ˜Ÿๅฐ˜๏ผ‰


Docs ArXiv X Discord Resources

๐ŸŽฏ +43.70% Accuracy vs. OpenAI Memory
๐Ÿ† Top-tier Long-term Memory + Personalization
๐Ÿ’ฐ Saves 35.24% Memory Tokens
LoCoMo 75.80 โ€ข LongMemEval +40.43% โ€ข PrefEval-10 +2568% โ€ข PersonaMem +40.75%

๐Ÿง  MemOS Plugin: Persistent Memory for Your AI Agents โœจ

MemOS Plugin Banner

Your lobsters and Hermes Agents now have the best memory system โ€” choose Cloud Service or Self-hosted to get started ๐Ÿƒ๐Ÿป

๐Ÿ”Œ Plugin
๐Ÿ’ก Core Features
๐Ÿงฉ Resources
๐Ÿง  memos-local-plugin 2.0
  • One local-first memory core for Hermes Agent and OpenClaw.
  • Self-evolving memory: L1 trace, L2 policy, L3 world model,
    and crystallized Skills driven by feedback.
๐ŸŒ Website ยท ๐Ÿ“– Docs ยท ๐Ÿ™ GitHub ยท ๐Ÿ“ฆ NPM
โ˜๏ธ OpenClaw Cloud Plugin๐Ÿ–ฅ๏ธ MemOS Dashboard ยท ๐Ÿ“– Full Tutorial

๐Ÿ‘พ MemOS: Memory Operating System for LLM & AI Agents

MemOS is a Memory Operating System for LLMs and AI agents that unifies store / retrieve / manage for long-term memory, enabling context-aware and personalized interactions with KB, multi-modal, tool memory, and enterprise-grade optimizations built in.

Key Features

  • Unified Memory API: A single API to add, retrieve, edit, and delete memoryโ€”structured as a graph, inspectable and editable by design, not a black-box embedding store.
  • Multi-Modal Memory: Natively supports text, images, tool traces, and personas, retrieved and reasoned together in one memory system.
  • Multi-Cube Knowledge Base Management: Manage multiple knowledge bases as composable memory cubes, enabling isolation, controlled sharing, and dynamic composition across users, projects, and agents.
  • Asynchronous Ingestion via MemScheduler: Run memory operations asynchronously with millisecond-level latency for production stability under high concurrency.
  • Memory Feedback & Correction: Refine memory with natural-language feedbackโ€”correcting, supplementing, or replacing existing memories over time.

News

  • 2026-05-09 ยท ๐Ÿง  memos-local-plugin 2.0 Official local memory plugin for Hermes Agent and OpenClaw. One core powers self-evolving memory across L1 traces, L2 policies, L3 world models, and crystallized Skills, with local-first storage and feedback-driven retrieval.

  • 2026-04-10 ยท ๐Ÿ‘ง๐Ÿป MemOS Hermes Agent Local Plugin Official Hermes Agent memory plugins launched: Hybrid retrieval (FTS5 + vector), smart dedup, tiered skill evolution, multi-agent collaboration. 100% local, zero cloud dependency.

  • 2026-03-08 ยท ๐Ÿฆž MemOS OpenClaw Plugin โ€” Cloud & Local Official OpenClaw memory plugins launched. Cloud Plugin: hosted memory service with 72% lower token usage and multi-agent memory sharing (MemOS-Cloud-OpenClaw-Plugin). Local Plugin (v1.0.0): 100% on-device memory with persistent SQLite, hybrid search (FTS5 + vector), task summarization & skill evolution, multi-agent collaboration, and a full Memory Viewer dashboard.

  • 2025-12-24 ยท ๐ŸŽ‰ MemOS v2.0: Stardust (ๆ˜Ÿๅฐ˜) Release Comprehensive KB (doc/URL parsing + cross-project sharing), memory feedback & precise deletion, multi-modal memory (images/charts), tool memory for agent planning, Redis Streams scheduling + DB optimizations, streaming/non-streaming chat, MCP upgrade, and lightweight quick/full deployment.

    โœจ New Features

    Knowledge Base & Memory

    • Added knowledge base support for long-term memory from documents and URLs

    Feedback & Memory Management

    • Added natural language feedback and correction for memories
    • Added memory deletion API by memory ID
    • Added MCP support for memory deletion and feedback

    Conversation & Retrieval

    • Added chat API with memory-aware retrieval
    • Added memory filtering with custom tags (Cloud & Open Source)

    Multimodal & Tool Memory

    • Added tool memory for tool usage history
    • Added image memory support for conversations and documents
    ๐Ÿ“ˆ Improvements

    Data & Infrastructure

    • Upgraded database for better stability and performance

    Scheduler

    • Rebuilt task scheduler with Redis Streams and queue isolation
    • Added task priority, auto-recovery, and quota-based scheduling

    Deployment & Engineering

    • Added lightweight deployment with quick and full modes
    ๐Ÿž Bug Fixes

    Memory Scheduling & Updates

    • Fixed legacy scheduling API to ensure correct memory isolation
    • Fixed memory update logging to show new memories correctly
  • 2025-08-07 ยท ๐ŸŽ‰ MemOS v1.0.0 (MemCube) Release First MemCube release with a word-game demo, LongMemEval evaluation, BochaAISearchRetriever integration, improved search capabilities, and the official Playground launch.

    โœจ New Features

    Playground

    • Expanded Playground features and algorithm performance.

    MemCube Construction

    • Added a text game demo based on the MemCube novel.

    Extended Evaluation Set

    • Added LongMemEval evaluation results and scripts.
    ๐Ÿ“ˆ Improvements

    Plaintext Memory

    • Integrated internet search with Bocha.
    • Expanded graph database support.
    • Added contextual understanding for the tree-structured plaintext memory search interface.
    ๐Ÿž Bug Fixes

    KV Cache Concatenation

    • Fixed the concat_cache method.

    Plaintext Memory

    • Fixed graph search-related issues.
  • 2025-07-07 ยท ๐ŸŽ‰ MemOS v1.0: Stellar (ๆ˜Ÿๆฒณ) Preview Release A SOTA Memory OS for LLMs is now open-sourced.

  • 2025-07-04 ยท ๐ŸŽ‰ MemOS Paper Release MemOS: A Memory OS for AI System is available on arXiv.

  • 2024-07-04 ยท ๐ŸŽ‰ Memory3 Model Release at WAIC 2024 The Memory3 model, featuring a memory-layered architecture, was unveiled at the 2024 World Artificial Intelligence Conference.


๐Ÿš€ Quick-start Guide

โ˜๏ธ 1ใ€Cloud API (Hosted)

Get API Key

Next Steps

๐Ÿ–ฅ๏ธ 2ใ€Self-Hosted (Local/Private)

  1. Get the repository.
    git clone https://github.com/MemTensor/MemOS.git
    cd MemOS
    pip install -r ./docker/requirements.txt
    
  2. Configure docker/.env.example and copy to MemOS/.env
  • The OPENAI_API_KEY,MOS_EMBEDDER_API_KEY,MEMRADER_API_KEY and others can be applied for through BaiLian.
  • Fill in the corresponding configuration in the MemOS/.env file.
  • Supported LLM providers: OpenAI, Azure OpenAI, Qwen (DashScope), DeepSeek, MiniMax, Ollama, HuggingFace, vLLM. Set MOS_CHAT_MODEL_PROVIDER to select the backend (e.g., openai, qwen, deepseek, minimax).
  1. Start the service.

Basic Usage (Self-Hosted)

  • Add User Message
    import requests
    import json
    
    data = {
        "user_id": "8736b16e-1d20-4163-980b-a5063c3facdc",
        "mem_cube_id": "b32d0977-435d-4828-a86f-4f47f8b55bca",
        "messages": [
            {
                "role": "user",
                "content": "I like strawberry"
            }
        ],
        "async_mode": "sync"
    }
    headers = {
        "Content-Type": "application/json"
    }
    url = "http://localhost:8000/product/add"
    
    res = requests.post(url=url, headers=headers, data=json.dumps(data))
    print(f"result: {res.json()}")
    
  • Search User Memory
    import requests
    import json
    
    data = {
        "query": "What do I like",
        "user_id": "8736b16e-1d20-4163-980b-a5063c3facdc",
        "mem_cube_id": "b32d0977-435d-4828-a86f-4f47f8b55bca"
    }
    headers = {
        "Content-Type": "application/json"
    }
    url = "http://localhost:8000/product/search"
    
    res = requests.post(url=url, headers=headers, data=json.dumps(data))
    print(f"result: {res.json()}")
    

๐Ÿ“š Resources

  • Awesome-AI-Memory This is a curated repository dedicated to resources on memory and memory systems for large language models. It systematically collects relevant research papers, frameworks, tools, and practical insights. The repository aims to organize and present the rapidly evolving research landscape of LLM memory, bridging multiple research directions including natural language processing, information retrieval, agentic systems, and cognitive science.
    Get started ๐Ÿ‘‰๐Ÿป IAAR-Shanghai/Awesome-AI-Memory

  • MemOS Cloud OpenClaw Plugin Official OpenClaw lifecycle plugin for MemOS Cloud. It automatically recalls context from MemOS before the agent starts and saves the conversation back to MemOS after the agent finishes.
    Get started ๐Ÿ‘‰๐Ÿป MemTensor/MemOS-Cloud-OpenClaw-Plugin


๐Ÿ’ฌ Community & Support

Join our community to ask questions, share your projects, and connect with other developers.

  • GitHub Issues: Report bugs or request features in our GitHub Issues.
  • GitHub Pull Requests: Contribute code improvements via Pull Requests.
  • GitHub Discussions: Participate in our GitHub Discussions to ask questions or share ideas.
  • Discord: Join our Discord Server.
  • WeChat: Scan the QR code to join our WeChat group.
QR Code

๐Ÿ“œ Citation

Note

We publicly released the Short Version on May 28, 2025, making it the earliest work to propose the concept of a Memory Operating System for LLMs.

If you use MemOS in your research, we would appreciate citations to our papers.


@article{li2025memos_long,
  title={MemOS: A Memory OS for AI System},
  author={Li, Zhiyu and Song, Shichao and Xi, Chenyang and Wang, Hanyu and Tang, Chen and Niu, Simin and Chen, Ding and Yang, Jiawei and Li, Chunyu and Yu, Qingchen and Zhao, Jihao and Wang, Yezhaohui and Liu, Peng and Lin, Zehao and Wang, Pengyuan and Huo, Jiahao and Chen, Tianyi and Chen, Kai and Li, Kehang and Tao, Zhen and Ren, Junpeng and Lai, Huayi and Wu, Hao and Tang, Bo and Wang, Zhenren and Fan, Zhaoxin and Zhang, Ningyu and Zhang, Linfeng and Yan, Junchi and Yang, Mingchuan and Xu, Tong and Xu, Wei and Chen, Huajun and Wang, Haofeng and Yang, Hongkang and Zhang, Wentao and Xu, Zhi-Qin John and Chen, Siheng and Xiong, Feiyu},
  journal={arXiv preprint arXiv:2507.03724},
  year={2025},
  url={https://arxiv.org/abs/2507.03724}
}

@article{li2025memos_short,
  title={MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models},
  author={Li, Zhiyu and Song, Shichao and Wang, Hanyu and Niu, Simin and Chen, Ding and Yang, Jiawei and Xi, Chenyang and Lai, Huayi and Zhao, Jihao and Wang, Yezhaohui and others},
  journal={arXiv preprint arXiv:2505.22101},
  year={2025},
  url={https://arxiv.org/abs/2505.22101}
}

@article{yang2024memory3,
author = {Yang, Hongkang and Zehao, Lin and Wenjin, Wang and Wu, Hao and Zhiyu, Li and Tang, Bo and Wenqiang, Wei and Wang, Jinbo and Zeyun, Tang and Song, Shichao and Xi, Chenyang and Yu, Yu and Kai, Chen and Xiong, Feiyu and Tang, Linpeng and Weinan, E},
title = {Memory$^3$: Language Modeling with Explicit Memory},
journal = {Journal of Machine Learning},
year = {2024},
volume = {3},
number = {3},
pages = {300--346},
issn = {2790-2048},
doi = {https://doi.org/10.4208/jml.240708},
url = {https://global-sci.com/article/91443/memory3-language-modeling-with-explicit-memory}
}

๐Ÿ™Œ Contributing

We welcome contributions from the community! Please read our contribution guidelines to get started.


๐Ÿ“„ License

MemOS is licensed under the Apache 2.0 License.