AI & HPC Data Centers
Fault Tolerant Solutions
Integrated Memory
Overcome traditional computational limitations, delivering unprecedented performance, scalability, and cost-effectiveness using CXL memory technology.
The memory wall is a particularly significant challenge in memory-intensive applications like artificial intelligence where training large models requires ultra-fast memory bandwidth which cannot keep up with the increasing demand.
The time it takes to move data between the GPU and memory (or across multiple GPUs) can become a significant bottleneck, slowing down training time.
Even for inference using trained models, the memory wall can lead to increased latency, as the model needs to access data from memory to make predictions.
If the memory system cannot keep up with the demands of inference requests, the overall throughput of the AI system can be reduced.
Scaling AI models to serve a large number of users can be difficult due to the memory limitations, requiring more hardware and complex infrastructure.
Along with other industry leaders, such as Alibaba, Cisco, Dell EMC, Facebook, Google, Hewlett Packard Enterprise, Intel Corporation and Microsoft, we have teamed up to form an open industry standard group to develop technical specifications that facilitate breakthrough performance for emerging usage models while supporting an open ecosystem for data center accelerators and other high-speed enhancements.
Compute Express Link® (CXL) is an industry, open standard protocol that redefines how servers manage memory and compute resources. By enabling high-speed, low-latency connections between CPUs and memory, CXL eliminates traditional bottlenecks in data processing, unlocking new levels of scalability and performance for data-intensive workloads which are increasingly used in emerging applications, such as artificial intelligence (AI), high-performance computing (HPC), and machine learning (ML).
For organizations requiring competitive insights faster, and speed and accuracy drive competitive advantage, CXL delivers game-changing benefits:
• Faster Data Processing: Real-time analysis of massive datasets with minimal delay.
• Improved Infrastructure Efficiency: Optimize resource utilization while reducing operational costs.
• Scalable, Future-Proof Solutions: Seamlessly expand memory to meet evolving data demands without costly infrastructure overhauls.
Our new family of Add-In Cards (AICs) which implements the CXL standard and also supports industry standard DDR5 DIMMs, are the first in their class of high-density DIMM AICs to adopt the CXL protocol. Our 4-DIMM and 8-DIMM products enable server and data center architects to quickly add up to 4TB of memory in a familiar, easy-to-deploy form factor.
AICs enable servers to reach up to 1TB of memory per CPU with cost effective 64GB RDIMMs. They also offer an opportunity for supply chain optionality. Replacing high density RDIMMs with a greater number of lower density modules can enable lower system memory costs depending on market conditions.
While AI, HPC, and ML are requiring larger amounts of high-speed memory that exceeds what current servers can accommodate, attempts to add more memory via the traditional DIMM-based parallel bus interface is becoming problematic due to pin limitations on CPUs.
With the the industry turning to CXL-based solutions which are more pin efficient, our 4-DIMM and 8-DIMM AICs are built using advanced CXL controllers which eliminate memory bandwidth bottlenecks and capacity constraints for emerging compute-intensive workloads.
Reach out to Penguin Solutions today to learn more about our CXL products and how we can help you leverage your data, turning information into actionable insights faster.
Reach out today and learn more how we can help you maximize your memory expansion and pooling capabilities while enabling lower cost scaling of memory capacity.