Skip to content
Tutorial emka
Menu
  • Home
  • Debian Linux
  • Ubuntu Linux
  • Red Hat Linux
Menu
deepseek engram

What is DeepSeek’s Engram?

Posted on January 14, 2026

DeepSeek has released a new technical paper detailing “Engram,” a conditional memory-based technique that allows AI models to utilize a queryable database of information committed to system memory. By committing sequences of data to static memory, Engram achieves demonstrably higher performance in long-context queries. This approach eases the reliance on reasoning for AI models, allowing GPUs to focus on more complex tasks. Crucially, this method increases performance while reducing the industry’s heavy reliance on scarce High-Bandwidth Memory (HBM).

The paper details how N-grams—statistical sequences of words—are integrated into the model’s neural networks, effectively placing them into a queryable memory bank. Engram allows models to simply “remember” facts rather than having to reason them out, a process that is far more computationally expensive. Released on the company’s GitHub page, Engram aims to curb the reliance on complex GPU memory by committing a knowledge library to more common system memory standards, such as CXL, enabling static memory to be held separately from an LLM’s compute power.

As detailed in the paper, an Engram-based model scaled to nearly 27 billion parameters can outperform a standard Mixture of Experts (MoE) model in long-context training. Standard MoE models utilize “conditional computation,” forcing the model to reconstruct data pieces every time they are referenced. Engram eliminates this computational waste by asking, “Do I already have this data?” This avoids what the paper describes as “expensive runtime reconstruction of a static lookup table,” saving valuable sequential depth for higher-level reasoning.

Engram is distinct from solutions like Nvidia’s KVCache, which offloads context data to NVMe memory. While KVCache acts as a short-term solution for remembering recent conversation history—akin to storing handwritten notes—Engram acts as a persistent record of a whole encyclopedia. Through tokenizer compression and “Multi-Head Hashing,” Engram reduces vocabulary size and allows for rapid parsing of information, ensuring distinct concepts (like “Universal” vs. “Universal Studios”) are retrieved without error via context-aware gating.

DeepSeek also explored the optimal balance between memory and compute, discovering a “U-curve” where allocating roughly 20–25% of the sparse parameter budget to Engram yields the best performance. In an experiment dubbed the “Infinite Memory Regime,” they found that performance scales linearly with memory size even when the compute budget is fixed. This implies that future AI improvements may not be solely bound by compute power, but could be achieved by expanding long-term “Engram” memory banks using standard DRAM within data centers.

The performance results highlight the potential of this architecture. In parallel testing, an Engram-27B model surpassed a standard 27B MoE model by 3.4 to 4 points in knowledge-intensive tasks and saw a massive leap in “Needle in a Haystack” long-context accuracy, scoring 97% compared to the MoE’s 84.2%. With DeepSeek viewing conditional memory as an “indispensable modeling primitive,” industry observers suggest this technology could be central to the rumored DeepSeek V4, potentially shifting hardware demand from HBM to standard system DRAM.

source: https://github.com/deepseek-ai/Engram

Recent Posts

  • Transmission 4.1.1 Released with Faster Performance and Better Efficiency for Linux and macOS
  • Calibre 9.3 Released with Enhances KEPUB Support for E-Book Readers
  • Microsoft Retires Outlook Contact Masking Feature March 31, 2026
  • Google Maps Privacy Update: Why Reviews and Ratings Are Now Hidden for Unlogged Users
  • Ubuntu 26.04 New Modular Firmware System: Faster Updates and Reduced System Bloat
  • Ghostty Linux Scrollbars Finally Here: What You Need to Know
  • Why Windows 11 Canary Channel Split into Two Builds? Explained!
  • What is Claude Cowork? And How Claude Cowork Uses Agentic AI
  • PocketBlue and Red Hat Bring Fedora Atomic Linux to Mobile Devices
  • Mozilla Ends Firefox Support for Windows 7, 8, and 8.1: What You Need to Know
  • Cosmic Desktop 1.0.7 Enhances Workspace Management: What’s New?
  • KDE Plasma 6.6 Released: What’s New and How to Upgrade?
  • Nginx Proxy Manager 2.14 Removes ARMv7 Support: What Users Need to Know
  •  KDE Plasma 6.6: A Complete Guide to the Latest Linux Desktop Features
  • Ubuntu 26.04 Resolute: Features, Release Date, and Everything You Need to Know
  • How to Fix Steam File Validation Error: Easy Steps for Beginners
  • 5 Essential PC Maintenance Tips to Keep Your Computer Fast and Healthy
  • What is Logseq? Forget Standard Notes App, Use this to Boosts Real Productivity
  • LibreOffice 25.8.5 Released with 62 Bug Fixes: What’s New?
  • Oracle’s New Plan for MySQL Community Engagement Explained
  • PipeWire 1.6 Brings LDAC Support and 128-Channel Audio: What’s New?
  • How to Fix Roblox Error: Create Support Files to Solve the Problem
  • Why Segmenting Your Home Network with VLANs Is the Upgrade You Didn’t Know You Needed
  • Proxmox 2026 Has The Best Backup and Recovery Feature
  • How to Calibrate Temperature and Humidity Sensors for Maximum Accuracy
  • Beda BRIVA dan Rekening? Ini Penjelasannya!
  • Pahami Perbedaan Kode SIEX, SIPX, dan SISX dengan Mudah!
  • Arti SPT Sebelumnya Tidak Ada dari BPS yang Perlu Kalian Pahami
  • Kode Error 205 di BCA Mobile: Penyebab dan Solusi Lengkap
  • Solusi Cepat Saat Voucher Axis Tidak Bisa Diproses
  • Prompt AI Menyusun Script Pola Suara Karakter agar Brand Jadi Ikonik
  • Prompt AI untuk Merancang Karakter Brand yang Ikonik
  • Prompt AI Audit Konten Sesuai Karakter Brand
  • Prompt AI Merubah Postingan LinkedIn Jadi Ladang Diskusi dengan ChatGPT
  • Prompt AI: Paksa Algoritma LinkedIn Promosikan Konten Kalian
  • Apa itu Spear-Phishing via npm? Ini Pengertian dan Cara Kerjanya yang Makin Licin
  • Apa Itu Predator Spyware? Ini Pengertian dan Kontroversi Penghapusan Sanksinya
  • Mengenal Apa itu TONESHELL: Backdoor Berbahaya dari Kelompok Mustang Panda
  • Siapa itu Kelompok Hacker Silver Fox?
  • Apa itu CVE-2025-52691 SmarterMail? Celah Keamanan Paling Berbahaya Tahun 2025
©2026 Tutorial emka | Design: Newspaperly WordPress Theme