Главная
Study mode:
on
1
Near Memory Compute for AI Inferencing
Description:
Explore a 21-minute technical presentation that delves into innovative solutions for AI inferencing data center challenges, focusing on Near-Memory Compute (NMC) technology. Learn how low-cost remote memory connected through low-latency interconnects can optimize inferencing operations. Discover the benefits of offloading specific inferencing tasks to smaller cores positioned near remote memory, supported by simulation data demonstrating reduced execution latency. Understand the potential impact on Total Cost of Ownership (TCO) reduction in inferencing data centers through the implementation of cost-effective remote memory pools. Gain insights into forward-thinking data center design principles that prioritize both sustainability and operational efficiency.

Near Memory Compute for AI Inferencing - Optimizing Data Center Design and TCO

Open Compute Project
Add to list