At #OCPSummit25 in San Jose, I sat down with Ventiva CEO Carl Schlachte to explore a counterintuitive idea: lessons from highly constrained laptop thermal envelopes can help stabilize AI servers and racks. We cover hotspot behavior under mixed CPU/GPU loads, coexistence with direct-to-chip liquid, and why “inside-the-box” gains still matter as power density soars. Worth a read if you’re balancing acoustics, reliability, and throughput at scale. Check out the full article: Link in comments. #Ventiva Open Compute Project Foundation
Ventiva CEO on thermal management for AI servers
More Relevant Posts
-
Avalue BMX-P820 – Power, Silence, and Flexibility in One Intel® 14th Gen Raptor Lake Core™ i7/i5/i3 (35W) CPU BMX-P820, the barebone system architecture gives customers full control over hardware configuration (CPU, DRAM, HDD/SSD) to build solutions tailored to their exact needs. With noise levels kept below 40dB, it provides a quiet yet powerful computing environment ideal for AI, automation, and edge computing. 🔗 Learn more: https://lnkd.in/gmBiVprQ
To view or add a comment, sign in
-
-
Latency, batch size, and throughput - what a combination! Lenovo's newest paper clarifies AI inferencing on servers powered by Intel CPUs, assisting companies in selecting the appropriate hardware for practical LLM tasks. With valuable information on performance optimization and cost analyses, this is essential reading for anyone trying to balance their AI goals with budget constraints. A hint: the ThinkSystem SR650 V3 exceeds expectations. Discover more by reading the blog: http://spr.ly/6045A7OUf | #WeAreLenovo
To view or add a comment, sign in
-
-
Celluster Reflex Compute continues to evolve quietly, refining the principles behind semantic scaling and intent-driven infrastructure. Instead of reacting to CPU or RAM thresholds, Reflex scales through awareness: secure boundaries, localized state, and lineage-based coordination. SDK preview planned for 2026 “Reflex doesn’t scale like containers rather it scales like thought.” From Cell to Celluster, from intent to identity, from execution to reflection, every primitive carries a piece of the story. Patent filed • all shared material is conceptual and illustrative. #ReflexCompute #Celluster #SemanticInfrastructure #AIInfrastructure #MovementGrade
To view or add a comment, sign in
-
-
This project provides a lightweight GPU management daemon for Kubernetes, handling device discovery, health checks, and metrics export for GPUs within nodes. More: https://ku.bz/9d6GVhXZD
To view or add a comment, sign in
-
-
We just released our paper https://lnkd.in/g7JuMdM9 on how Meta scaled GPU communication to 100k+ GPUs. The open source implementation is a part of the torchcomms repo we announced last week https://lnkd.in/gpqke9n4 including our implementations of various concepts like DQPLB, AllToAllvDynamic and more
To view or add a comment, sign in
-
-
Introducing the new VCM-2000A — a rugged computing platform offering scalable performance from Intel® Core™ i3 to Core™ i9 processors, designed to operate reliably in extended temperatures ranging from −40°C to +60°C.
To view or add a comment, sign in
-
The CPU landscape keeps shifting. AMD's Zen 6 promises a big IPC boost, AI features with NPUs, and a 2nm process. Great for professionals seeking higher performance with demanding AI workloads. Intel-NVIDIA's new alliance brings exciting x86-RTX SoCs, promising high CPU-GPU bandwidth. Choice matters in meeting your specific workflow needs. Design a custom workstation that maximizes these new advancements and fits your budget. Visit https://buildpc.net/ to get started.
To view or add a comment, sign in
-
-
“Honey, I shrunk the supercomputer” China unveils the world’s first mini AI server that’s the size of a small fridge, but has the capabilities of a room-sized supercomputer! BI Explorer computing system, or BIE-1, has 1,152 CPUs, 4.8 TB of DDR5 memory and 204 TB of storage! It’s from Guangdong Institute of Intelligence Science and Technology (GDIIST). Source : https://lnkd.in/dWEur3fH
To view or add a comment, sign in
-
-
The cats are getting out of the lab and into SLURM! 🐈⬛🚀 We’ve teamed up with the STFC Hartree Centre to plug our cat qubits into the world’s main HPC workflow tool, SLURM. ⚙ If you use a supercomputer, you’ll know SLURM. It runs on over 60% of the world’s supercomputers, slotting tasks onto the right CPUs, GPUs, and even QPUs. 🧠 This collaboration lays the bricks to bring quantum compute into those same workflows. ⚡💻 For the first time, cat-qubit processors will be able to operate inside an HPC environment as just another compute resource, a necessary step in getting quantum ready for practical, real-world use. 🔬🚀 Today only a handful of quantum hardware teams worldwide are working to integrate with SLURM, and our cats are proudly one of them. The early fault-tolerant era will unlock applications that classical computers alone could never do. But since these applications will be executed as a choir rather than a solo, we need true hybrid systems, where quantum and classical compute work side by side, each playing to their strengths. ⚛️🤝💻 Big thanks to the Hartree Centre for working with us on this step toward seamless quantum–classical integration! 🙌 #QuantumComputing #HPC #SLURM #CatQubits
To view or add a comment, sign in
-
-
See these charts. If/when geopolitics or supply breaks the GPU flow, small fine-tuned models win—compute-lean, deployable on CPUs or modest GPUs. Parameter count is debt; latency, energy, and locality are assets. Building specialized, edge-first inference is probably not a bad idea. Then compute shock becomes a moat, not an outage.
To view or add a comment, sign in
-
https://techarena.ai/content/from-laptops-to-data-center-racks-ventivas-thermal-play-at-ocp