
EverOS
2
models
1
languages
Description
EverOS is a long-term memory operating system for AI agents, developed by EverMind AI (San Mateo, CA). It addresses the architectural limitation of stateless large language models by providing a structured memory infrastructure that persists, consolidates, and retrieves knowledge across sessions. The platform is available both as an open-source self-hosted deployment (GitHub) and as EverOS Cloud at everos.evermind.ai.
Architecture
EverOS operates through a four-layer architecture: the Agentic Layer (task understanding, planning, execution), the Memory Layer (long-term storage and retrieval), the Index Layer (embeddings, key-value pairs, knowledge graph indexing), and the API/MCP Interface Layer (integration with external enterprise systems). Memory management follows an engram-inspired three-phase lifecycle: Episodic Trace Formation (converting dialogue streams into structured MemCells), Semantic Consolidation (organizing MemCells into thematic MemScenes), and Reconstructive Recollection (MemScene-guided agentic retrieval composing necessary context for downstream reasoning).
Core Innovations
EverOS features four key innovations: (1) Self-Evolving Agent Memory β a Skills Evolution Engine that automatically distills reusable skills (SOPs) from completed tasks via agent case extraction, semantic clustering, and skill emergence; (2) mRAG Hybrid Retrieval Architecture β native multimodal memory ingestion (PDFs, images, Word docs, spreadsheets, URLs) through a single API, fusing dense semantic vectors, sparse keyword matching, and multimodal alignment; (3) HyperMem Architecture β a hypergraph memory network (accepted at ACL 2026) replacing flat vector databases to capture multi-hop, cross-temporal entity relationships with ultra-low latency; (4) RESTful API with MCP interface and EverOS Cloud Playgrounds (Coding Playground integrated with Google Colab, Chat Playground for visual memory comparison).
Benchmark Performance
EverCore (the underlying engine) achieves 93.05% accuracy on LoCoMo, 83.00% on LongMemEval, 90.04% on HaluMem, and SOTA performance on PersonaMem v2. The Skills Evolution Engine yielded a 234.8% relative increase in task success rate for software engineering problems (27B model, EvoAgentBench). The underlying research is published as arXiv:2601.02163 (EverMemOS).
Storage & Integration
Supported storage engines include SQLite, PostgreSQL, and vector databases with embeddings (FAISS, Milvus, pgvector). The platform supports multi-tenant memory IDs and is compatible with LangGraph, Haystack, and other agent frameworks. It integrates with any LLM API (OpenAI, Qwen, Llama, local models via API wrapper) and embedding services. All memories are stored as transparent JSON objects with timestamped, persistent episodes.
Use Cases
EverOS is designed for personalized AI assistants, customer service agents requiring continuous contextual understanding, multi-user collaboration and knowledge retention, research and analysis (building knowledge bases from conversations), and educational tools adapting to learning patterns.
MLOps / LLMOps Lifecycle
- Artifact versioning
- Approval workflows
- Immutable artifacts
- Lineage tracking
- Online serving (low-latency access)
- Offline storage (historical training)
- Streaming ingestion
- Prompt registry
- Versioning
- Testing frameworks
- Data drift detection
- Concept drift detection
- Hallucination monitoring
- Bias evaluation tools
- Labeling services
- RLHF workflows
- Manual override mechanisms
Data & Knowledge
Security
Developer Ecosystem
Pricing & Business Model
Pricing models
Resource quotas
SLA & Support