By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
TechgoonduTechgoonduTechgoondu
  • Audio-visual
  • Enterprise
    • Software
    • Cybersecurity
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
    • Cellphones
    • Tablets
  • PC
  • Telecom
Search
© 2023 Goondu Media Pte Ltd. All Rights Reserved.
Reading: Computex 2025: A glimpse of the machines inside Nvidia’s “AI factories”
Share
Font ResizerAa
TechgoonduTechgoondu
Font ResizerAa
  • Audio-visual
  • Enterprise
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
  • PC
  • Telecom
Search
  • Audio-visual
  • Enterprise
    • Software
    • Cybersecurity
  • Gaming
  • Imaging
  • Internet
  • Media
  • Mobile
    • Cellphones
    • Tablets
  • PC
  • Telecom
Follow US
© 2023 Goondu Media Pte Ltd. All Rights Reserved.
Techgoondu > Blog > Enterprise > Computex 2025: A glimpse of the machines inside Nvidia’s “AI factories”
Enterprise

Computex 2025: A glimpse of the machines inside Nvidia’s “AI factories”

Alfred Siew
Last updated: May 22, 2025 at 1:43 AM
Alfred Siew
Published: May 22, 2025
10 Min Read
SHARE

If you believe AI proponents like Nvidia head honcho Jensen Huang, data centres will become newfangled “AI factories” that pump out “intelligence” or insights to help people run their everyday lives, say, learn skills faster or run a business more efficiently.

Instead of making shoes, generating electricity or sending Internet data to you, such AI factories will deliver a much-needed “intelligence” that changes everything, from powering a chatbot to a physical robot in future.

And like any factory, these AI factories need a proper bunch of machines to produce this intelligence. This is where Nvidia’s much-sought-after chips dominate the headlines and even engender great-power rivalry today.

AI-generated image of an AI factory. ILLUSTRATION: Nvidia

At the Computex show this week, we got a close look at some Nvidia-powered AI servers that are doing all the heavy lifting to make AI possible.

These tasks include training large language models (so they become smart), inferencing from the trained knowledge (to sound smart in a response) and developing an an app or service (to use AI for real, such as in scientific research).

Asus, one of Nvidia’s many partners that make the AI machines using its graphics processing units (GPUs), showed us a few interesting things in Taipei:

1. AI servers are super dense

Just a few years ago, a typical rack of servers in a data centre usually had a capacity of under 10kW. This was because they typically ran “simple” tasks such as websites, office productivity or payroll and accounting. These tasks have been well optimised over time.

Then came AI. To train an AI model, it takes a tremendous amount of computing power, say, to learn the patterns and recognise and differentiate one thing from another. The size of the data sets involved matter, too. Today, an AI data centre rack can be so densely packed with computing power that its capacity exceeds 100kW.

At the Asus booth at Computex, for example, we saw an Nvidia H200-based setup that sucks up to 74.4kW – that’s from eight nodes each drawing 9.3kW.

It’s not even Nvidia’s most powerful machine – the latest GB200- and GB300-based servers, which Asus also sells in a rack called AI Pod, are even more advanced. Needless to say, these machines pack even more compute in such a small enclosure that’s slotted into standard data centre racks.

2. Water cooling is preferred

The added computer power in a small enclosed space means these machines run very hot. Fans are no longer efficient enough and many data centre operators have turned to liquid cooling. It helps that liquid is a better conductor of heat than air – you spend less energy transferring that heat away.

Asus’ GB200- and GB300-based AI servers, for example, come in both air and liquid cooling options. For liquid cooling, a copper heat pipe takes the heat away from the hot-running chips. It is cooled down further by fans before the coolant is returned to cool down the chip again.

A liquid cooling setup with Nvidia Blackwell GPUs. PHOTO: Alfred Siew

Another even more exotic form of liquid cooling is immersion cooling. This means sliding entire electronic components – usually a chip and motherboard – into a non-conductive liquid to cool it down.

As you’d imagine, this is an even more efficient cooling method, though it’s also harder to follow. Asus doesn’t make immersion cooling kits but data centre equipment makers such as Vertiv do.

It costs extra money to set up these large tanks of liquid and you still need to reconfigure a data centre’s server halls – for one, you don’t need so many fans now.

And it’s somewhat ironic that liquid cooling – including immersion cooling – has long been a hobbyist pursuit for PC enthusiasts, who today have their own closed-loop liquid coolers for their CPUs.

Notably, in another hall in Computex this week, an even more exotic method – liquid nitrogen from large tanks – was used to cool down CPUs in a contest to see who could push their PCs to extremes.

3. They scale out and scale up

Okay, this is a little bit of Nvidia marketing-speak but it speaks to the different needs of businesses today. Depending on the AI capabilities they wish to focus on, they can either scale out with more and smaller Nvidia-based machines or scale up with more powerful Nvidia-based machines.

At the extreme end, Asus’ AI Pod comes in a entire rack that can fit 72 Nvidia Blackwell Ultra GPUs and 36 Grace CPUs. You can train complex AI models with 1 trillion parameters, the upper end of such activities today.

To “build up” to these capabilities, you’d need deep pockets, of course. Or buy time or tokens on one of these machines run by a hyperscale cloud provider.

This Asus server packs Nvidia’s latest Blackwell GPUs for heavy AI workloads. PHOTO: Alfred SIew

At the slightly lower end, Asus also offers smaller servers that pack in older but still powerful Nvidia H200 GPUs or the new RTX Pro 6000 Blackwell Server Edition chips that promise to great AI inference performance for AI agents.

These servers help to “scale out” AI capabilities by bringing them to more parts of the enterprise, enabling employees to take on the computing power in an AI factory to, say, generate images faster or deliver customer answers more coherently.

4. AI can be run on small machines

Don’t think everything AI needs to be run in data centre. Now, Nvidia and many of its technology partners even have lunchbox-sized PCs that deliver enough horsepower to test-drive, say, your latest research without having to go to the cloud, set things up, test and then scrub the data for security.

Asus’ AScent GX10, small AI computer seen at the Computex show in Taipei on May 20. PHOTO: Alfred Siew

The Ascent GX10 from Asus, for example, packs an Nvidia GB10 Grace Blackwell GPU capable of prototyping, fine-tuning, and inferring AI models with up to 200 billion parameters. And it is not much bigger than a mini PC.

A lesson from Sun Microsystems

It’s true Nvidia isn’t the only game in town when it comes to AI – AMD and Intel both offer alternatives, plus there are ways to use AI accelerators instead of full-fledged Nvidia GPUs for some tasks that don’t need the extreme performance.

How Nvidia has locked up the market is with its popular CUDA software. It enables developers to get up to speed with deep learning and other tasks with a ready-made set of tools.

Another factor is Nvidia’s high-speed NVLink connections between the various components in an AI server, which not only speed things up but also help join together multiple GPUs (like in the Asus AI Pod) to deliver a lot more performance.

Yet, Nvidia should take a lesson from Sun Microsystems? Who? Yes, the company that once called itself the dot in dotcom in the late 1990s and early 2000s, because it was the only one who could make servers powerful enough to enable large-scale websites for e-commerce.

Unfortunately, people later found a way to link up multiple slower but cheaper Intel servers to deliver good performance that rivalled the specialised and costly Sun machines.

This was the seed need to grow cloud computing to what it is today. It was built on such “commodity” servers from Intel and AMD. They replaced larger monolithic ones by Sun, which got sold to Oracle in 2010.

Just like Sun, Nvidia is selling the picks and shovels in yet another gold rush about 20 years later. It’s the biggest player in town, as Sun was. Now, the question is whether it can retain its lead by selling a vision of AI factories that depend on its chips deliver the gold.

Asia-Pacific PC market declines in the run-up to Windows 8, says IDC
When your company wants to place a corporate Wi-Fi access point in your home
Life goes digital after global lockdown to fight coronavirus
Blockchain technology needs a proof of concept in the real world
Q&A: Southeast Asian governments face advanced persistent threats online, says FireEye
TAGGED:AIAI chipAI FactoryasusComputex 2025GB300gpuGrace BlackwellH200Nvidia

Sign up for the TG newsletter

Never miss anything again. Get the latest news and analysis in your inbox.

By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Whatsapp Whatsapp LinkedIn Copy Link Print
Avatar photo
ByAlfred Siew
Follow:
Alfred is a writer, speaker and media instructor who has covered the telecom, media and technology scene for more than 20 years. Previously the technology correspondent for The Straits Times, he now edits the Techgoondu.com blog and runs his own technology and media consultancy.
Previous Article The three megatrends shaping society and technology
Leave a Comment

Leave a ReplyCancel reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Stay Connected

FacebookLike
XFollow

Latest News

The three megatrends shaping society and technology
Enterprise Internet
May 21, 2025
Computex 2025: Asus ROG Falcata impresses with ergonomic feel
Gaming PC
May 20, 2025
Computex 2025: To broaden AI footprint, Nvidia offers fast links to third-party chips
Enterprise
May 19, 2025
Promising speed and better coverage, Singtel 5G+ targets premium users
Mobile Telecom
May 15, 2025

Techgoondu.com is published by Goondu Media Pte Ltd, a company registered and based in Singapore.

.

Started in June 2008 by technology journalists and ex-journalists in Singapore who share a common love for all things geeky and digital, the site now includes segments on personal computing, enterprise IT and Internet culture.

banner banner
Everyday DIY
PC needs fixing? Get your hands on with the latest tech tips
READ ON
banner banner
Leaders Q&A
What tomorrow looks like to those at the leading edge today
FIND OUT
banner banner
Advertise with us
Discover unique access and impact with TG custom content
SHOW ME

 

 

POWERED BY READYSPACE
The Techgoondu website is powered by and managed by Readyspace Web Hosting.

TechgoonduTechgoondu
© 2024 Goondu Media Pte Ltd. All Rights Reserved | Privacy | Terms of Use | Advertise | About Us | Contact
Join Us!
Never miss anything again. Get the latest news and analysis in your inbox.

Zero spam, Unsubscribe at any time.
 

Loading Comments...
 

    Welcome Back!

    Sign in to your account

    Username or Email Address
    Password

    Lost your password?