• DeepSeek’s Engram separates static memory from computation, increasing efficiency in large AI models
  • The method reduces high-speed memory needs by enabling DeepSeek models to use lookups
  • Engram supports asynchronous prefetching across multiple GPUs with minimal performance overhead

DeepSeek, in collaboration with Peking University, introduced a new training method called Engram, designed to decouple memory storage from computational processes.

Traditional large language models require high-bandwidth memory for knowledge retrieval and basic computation, creating a bottleneck in both performance and cost.

GPU and system memory architectures, potentially avoiding costly HBM upgrades.

This technique may relieve pressure on expensive memory hardware, particularly in regions such as China, where HBM access lags behind competitors such as Samsung, SK Hynix, and Micron.

Early validation of Engram suggests models can expand parameter scale and reasoning capacity while managing memory demands more efficiently.

This approach may help ease memory constraints across AI infrastructure, potentially reducing sharp DDR5 DRAM price swings.

Via SCMP

Follow TechRadar on Google News and add us as a preferred source to get our expert news, reviews, and opinion in your feeds. Make sure to click the Follow button!

And of course you can also follow TechRadar on TikTok for news, reviews, unboxings in video form, and get regular updates from us on WhatsApp too.