Intelligence Begins with Memory

Build a world-leading open-source infrastructure for large-model-oriented memory

Community Projects

Open Collaboration, Sharing the Power of Memory
Memory Systems and Frameworks
  • MemOS Open-Source Memory Operating System
  • LightMem Lightweight Pluggable Memory System for Large Models
  • Text2Mem Unified Memory Operation Language
RAG and Doc Optimization
  • KaLM-V2 High-Performance Universal Text Encoder
  • MoM Document Memory Extraction for RAG
  • MoC Hybrid Document Chunking Expert for RAG
Benchmarks
  • HaluMem Memory Hallucination Evaluation Framework
  • RecCocktail Parameterized Memory Fusion Method for Personalized Recommendation Systems
  • SafeRAG RAG Security Evaluation Benchmark
  • CRUD-RAG Comprehensive Chinese RAG Evaluation Benchmark

Open-Source Memory Operating System

A native memory framework developed by the MemTensor team for building intelligent systems capable of remembering, adapting, and evolving.

LightMem

Lightweight Pluggable Memory System for LLM

Addressing common challenges faced by large models in long-term interactions—such as limited context, information forgetting, and memory redundancy—LightMem introduces a human-memory-inspired lightweight architecture. Through semantic compression and noise filtering, it significantly reduces redundant information; dynamic semantic segmentation improves memory organization and retrieval efficiency; and an offline “sleep-like” update mechanism enables memory reorganization and knowledge consolidation. LightMem maintains high memory fidelity and consistency while reducing computational and storage costs, providing an efficient and scalable long-term memory solution for large language models and intelligent agents, supporting personalized and complex application scenarios.

Text2Mem

Unified Memory Operation Language

Proposes a unified language, Text2Mem, for intelligent agent memory management. It converts natural language into standardized JSON commands covering various operations such as encoding, storage, and retrieval. Through parsing, validation, and adaptation layers, it enables secure, deterministic, and portable execution across heterogeneous memory backends, providing a standardized foundation for memory control.

KaLM-V2

High-Performance Universal Text Encoder

To address the balance challenge between retrieval efficiency and precision in RAG systems, a multi-scale universal embedding model is proposed. In the MTEB benchmark—which covers thousands of languages, hundreds of tasks, and nine major task types—it ranks first globally, surpassing major open- and closed-source models such as NVIDIA’s nvidia/llama-embed-nemotron-8b, Alibaba’s Qwen/Qwen3-Embedding-8B, and Google’s Google/gemini-embedding-001. It has been widely deployed and downloaded over one million times.

MoM

Document Memory Extraction for RAG

Upgrades traditional RAG’s passive chunking into proactive document memory construction. It leverages expert-level large models to generate logical outlines and key segments, combines multi-path sampling and multi-dimensional evaluation to select high-quality memories, and trains smaller models to acquire human-like reading and three-tier memory retrieval capabilities—enhancing multi-domain retrieval and generation performance.

MoC

Hybrid Document Chunking Expert for RAG

Introduces two new metrics—boundary clarity and block cohesion—for evaluating text segmentation quality, and builds a granularity-aware hybrid chunking framework. This guides models to generate regularized segmentation rules, producing structured, high-quality text chunks that significantly improve RAG retrieval and generation performance while maintaining a balance between computational efficiency and precision.

HaluMem

Memory Hallucination Evaluation Framework

Proposes the industry’s first hallucination evaluation framework designed for memory operating systems. It supports fine-grained and procedural hallucination assessment, helping memory systems quickly locate hallucination types and identify optimization directions.

RecCocktail

Parameterized Memory Fusion Method for Personalized Recommendation Systems

For personalized recommendation scenarios, this method adaptively fuses domain-general user parameterized memory with scenario-specific user parameterized memory in the weight space. Through a plug-and-play modular design, it enables efficient integration of any domain memory into core recommendation capabilities without additional inference overhead. It significantly improves personalized recommendation model performance across diverse application scenarios.

SafeRAG

RAG Security Evaluation Benchmark

SafeRAG SafeRAG introduces a security evaluation benchmark for RAG systems, constructing datasets that cover various attack tasks such as noise injection, context conflict, soft advertising, and denial-of-service. The benchmark systematically simulates multiple attack scenarios and empirically reveals critical vulnerabilities of existing RAG components when facing malicious knowledge manipulation.

CRUD-RAG

Comprehensive Chinese RAG Evaluation Benchmark

Builds a large-scale Chinese retrieval-augmented generation evaluation benchmark. It systematically designs datasets and metrics for four core application scenarios—Create, Read, Update, and Delete—jointly evaluating the performance of retrievers, knowledge bases, and large models. It provides empirical references for optimizing RAG systems under different real-world conditions.

Academic Achievements

Research-Driven, Inspiring Memory Intelligence
MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models
2025-5-28
HaluMem: Evaluating Hallucinations in Memory Systems of Agents
2025-11-5
LightMem: Lightweight and Efficient Memory-Augmented Generation
2025-10-21
A Survey on the Memory Mechanism of Large Language Model-based Agents
2025-9-10
CMT: A Memory Compression Method for Continual Knowledge Learning of Large Language Models
2024-12-10
KaLM-Embedding-V2: Superior Training Techniques and Data Inspire A Versatile Embedding Model
2025-6-26
RecCocktail: A Generalizable and Efficient Framework for LLM-Based Recommendation
2025-10-30
An Efficient LLM-based Evolutional Recommendation with Locate-Forget-Update Paradigm
2025-11-20
Text2Mem: A Unified Memory Operation Language for Memory Operating System
2025-9-14
MoM: Mixtures of Scenario-Aware Document Memories for Retrieval-Augmented Generation Systems
2025-10-16
SafeRAG: Benchmarking Security in Retrieval-Augmented Generation of Large Language Model
2025-1-28
MoC: Mixtures of Text Chunking Learners for Retrieval-Augmented Generation System
2025-3-12
CRUD-RAG: A Comprehensive Chinese Benchmark for Retrieval-Augmented Generation of Large Language Models
2024-1-30
Memory Decoder: A Pretrained, Plug-and-Play Memory for Large Language Models
2025-8-13
MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models
2025-5-28
HaluMem: Evaluating Hallucinations in Memory Systems of Agents
2025-11-5
LightMem: Lightweight and Efficient Memory-Augmented Generation
2025-10-21
A Survey on the Memory Mechanism of Large Language Model-based Agents
2025-9-10
CMT: A Memory Compression Method for Continual Knowledge Learning of Large Language Models
2024-12-10
KaLM-Embedding-V2: Superior Training Techniques and Data Inspire A Versatile Embedding Model
2025-6-26
RecCocktail: A Generalizable and Efficient Framework for LLM-Based Recommendation
2025-10-30
An Efficient LLM-based Evolutional Recommendation with Locate-Forget-Update Paradigm
2025-11-20
Text2Mem: A Unified Memory Operation Language for Memory Operating System
2025-9-14
MoM: Mixtures of Scenario-Aware Document Memories for Retrieval-Augmented Generation Systems
2025-10-16
SafeRAG: Benchmarking Security in Retrieval-Augmented Generation of Large Language Model
2025-1-28
MoC: Mixtures of Text Chunking Learners for Retrieval-Augmented Generation System
2025-3-12
CRUD-RAG: A Comprehensive Chinese Benchmark for Retrieval-Augmented Generation of Large Language Models
2024-1-30
Memory Decoder: A Pretrained, Plug-and-Play Memory for Large Language Models
2025-8-13
MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models
2025-5-28
HaluMem: Evaluating Hallucinations in Memory Systems of Agents
2025-11-5
LightMem: Lightweight and Efficient Memory-Augmented Generation
2025-10-21
A Survey on the Memory Mechanism of Large Language Model-based Agents
2025-9-10
CMT: A Memory Compression Method for Continual Knowledge Learning of Large Language Models
2024-12-10
KaLM-Embedding-V2: Superior Training Techniques and Data Inspire A Versatile Embedding Model
2025-6-26
RecCocktail: A Generalizable and Efficient Framework for LLM-Based Recommendation
2025-10-30
An Efficient LLM-based Evolutional Recommendation with Locate-Forget-Update Paradigm
2025-11-20
Text2Mem: A Unified Memory Operation Language for Memory Operating System
2025-9-14
MoM: Mixtures of Scenario-Aware Document Memories for Retrieval-Augmented Generation Systems
2025-10-16
SafeRAG: Benchmarking Security in Retrieval-Augmented Generation of Large Language Model
2025-1-28
MoC: Mixtures of Text Chunking Learners for Retrieval-Augmented Generation System
2025-3-12
CRUD-RAG: A Comprehensive Chinese Benchmark for Retrieval-Augmented Generation of Large Language Models
2024-1-30
Memory Decoder: A Pretrained, Plug-and-Play Memory for Large Language Models
2025-8-13
MemOS: An Operating System for Memory-Augmented Generation (MAG) in Large Language Models
2025-5-28
HaluMem: Evaluating Hallucinations in Memory Systems of Agents
2025-11-5
LightMem: Lightweight and Efficient Memory-Augmented Generation
2025-10-21
A Survey on the Memory Mechanism of Large Language Model-based Agents
2025-9-10
CMT: A Memory Compression Method for Continual Knowledge Learning of Large Language Models
2024-12-10
KaLM-Embedding-V2: Superior Training Techniques and Data Inspire A Versatile Embedding Model
2025-6-26
RecCocktail: A Generalizable and Efficient Framework for LLM-Based Recommendation
2025-10-30
An Efficient LLM-based Evolutional Recommendation with Locate-Forget-Update Paradigm
2025-11-20
Text2Mem: A Unified Memory Operation Language for Memory Operating System
2025-9-14
MoM: Mixtures of Scenario-Aware Document Memories for Retrieval-Augmented Generation Systems
2025-10-16
SafeRAG: Benchmarking Security in Retrieval-Augmented Generation of Large Language Model
2025-1-28
MoC: Mixtures of Text Chunking Learners for Retrieval-Augmented Generation System
2025-3-12
CRUD-RAG: A Comprehensive Chinese Benchmark for Retrieval-Augmented Generation of Large Language Models
2024-1-30
Memory Decoder: A Pretrained, Plug-and-Play Memory for Large Language Models
2025-8-13

Partners

Create and Win Together to Drive the Development of Memory Intelligence
Shanghai Jiao Tong University
Peking University
Zhejiang University
Tongji University
Renmin University of China
Beihang University
Nankai University
Fudan University
Harbin Institute of Technology
Harbin Engineering University
Shanghai University of Finance and Economics
Hefei University of Technology
Shanghai Jiao Tong University
Peking University
Zhejiang University
Tongji University
Renmin University of China
Beihang University
Nankai University
Fudan University
Harbin Institute of Technology
Harbin Engineering University
Shanghai University of Finance and Economics
Hefei University of Technology
Shanghai Jiao Tong University
Peking University
Zhejiang University
Tongji University
Renmin University of China
Beihang University
Nankai University
Fudan University
Harbin Institute of Technology
Harbin Engineering University
Shanghai University of Finance and Economics
Hefei University of Technology
Shanghai Jiao Tong University
Peking University
Zhejiang University
Tongji University
Renmin University of China
Beihang University
Nankai University
Fudan University
Harbin Institute of Technology
Harbin Engineering University
Shanghai University of Finance and Economics
Hefei University of Technology
IAAR
Memtensor
China Telecom
HAISUM
HIAS,UCAS
IAAR
Memtensor
China Telecom
HAISUM
HIAS,UCAS
IAAR
Memtensor
China Telecom
HAISUM
HIAS,UCAS
IAAR
Memtensor
China Telecom
HAISUM
HIAS,UCAS