1 min read

White Paper: Study of Jet Impinging and Integrated Cold Plate for Unleashing Chipset Power

White Paper: Study of Jet Impinging and Integrated Cold Plate for Unleashing Chipset Power
White Paper: Study of Jet Impinging and Integrated Cold Plate for Unleashing Chipset Power
1:28

As thermal design power (TDP) for modern processors such as CPUs, GPUs, and TPUs exceeds 1 kW, traditional air cooling methods are proving insufficient. Liquid cooling, particularly using cold plates, has emerged as a superior solution. This paper from Wiwynn introduces an advanced cold plate design that integrates in-package liquid cooling with jet impingement techniques to minimize junction-to-liquid thermal resistance, eliminating multiple layers of thermal interface material (TIM) and improving heat transfer efficiency.

The proposed integrated cold plate was evaluated through numerical simulations and experimental setups. Simulation results show that these integrated cold plates outperform traditional designs, reducing thermal resistance by 8%. Additional fluidic channels within the chip package, along with direct liquid cooling, further enhance performance by up to 25.7%. The jet impingement cold plate design significantly improves thermal resistance by 20–48% compared to traditional cold plates. When combined with the integrated cold plate, the impingement design reduces thermal resistance by up to 53%.

These advancements demonstrate that integrated impingement cold plates offer significant improvements in thermal management for high TDP processors, making them a promising technology for data centers and AI applications.

Register to Download the whitepaper!

 

White Paper: GPU-Initiated, Liquid-Cooled, Ultra-High-Density Storage for Next-Gen AI

1 min read

White Paper: GPU-Initiated, Liquid-Cooled, Ultra-High-Density Storage for Next-Gen AI

This paper introduces a paradigm shift in storage architecture designed to overcome the CPU-centric data path bottlenecks in modern AI workloads. By...

Read More
White Paper: KV Cache Offload to Improve AI Inferencing Cost and Performance

1 min read

White Paper: KV Cache Offload to Improve AI Inferencing Cost and Performance

This paper explores a disaggregated key-value (KV) storage architecture designed to efficiently offload KV cache tensors for generative AI workloads.

Read More
Autonomous AI Agent for End-to-End Component Data Extraction

1 min read

Autonomous AI Agent for End-to-End Component Data Extraction

This paper explores an advanced framework designed to automate the extraction of important attributes from unstructured part datasheets. By...

Read More