Sponsored by:

Visit AMD Visit Supermicro

Capture the full potential of IT

Vultr, Supermicro, AMD team to offer hi-performance cloud compute & AI infrastructure

Featured content

Vultr, Supermicro, AMD team to offer hi-performance cloud compute & AI infrastructure

Vultr, a global provider of cloud services, now offers Supermicro servers powered by AMD Instinct GPUs.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Supermicro servers powered by the latest AMD Instinct GPUs and supported by the AMD ROCm open software ecosystem are at the heart of a global cloud infrastructure program offered by Vultr.

Vultr calls itself a modern hyperscaler, meaning it provides cloud solutions for organizations facing complex AI and HPC workloads, high operational costs, vendor lock-in, and the need for rapid insights.

Launched in 2014, Vultr today offers services from 32 data centers worldwide, which it says can reach 90% of the world’s population in under 40 milliseconds. Vultr’s services include cloud instances, dedicated servers, cloud GPUs, and managed services for database, cloud storage and networking.

Vultr’s customers enjoy benefits that include costs 30% to 50% lower than those of the hyperscalers and 20% to 30% lower than those of other independent cloud providers. These customers—there are over 220,000 of them worldwide—also enjoy Vultr’s full native AI stack of compute, storage and networking.

Vultr is the flagship product of The Constant Co., based in West Palm Beach, Fla. The company was founded by David Aninowsky, an entrepreneur who also started GameServers.com and served as its CEO for 18 years.

Now Vultr counts among its partners AMD, which joined the Vultr Cloud Alliance, a partner program, just a year ago. In addition, AMD’s venture group co-led a funding round this past December that brought Vultr $333 million.

Expanded Data Center

Vultr is now expanding its relationship with Supermicro, in part because that company is first to market with the latest AMD Instinct GPUs. Vultr is now offering Supermcro systems powered by AMD Instinct MI355X, MI325X and MI300X GPUs. And as part of the partnership, Supermicro engineers work on-site with Vultr technicians.

Vultr is also relying on Supermicro for scaling. That’s a challenge for large AI implementations, as these configurations require deep expertise for both integration and operations.

Among Vultr’s offerings from Supermicro is a 4U liquid-cooled server (model AS -4126GS-NMR-LCC) with dual AMD EPYC 9005/9004 processors and up to eight AMD GPUs—the user’s choice of either MI325X or MI355X.

Another benefit of the new arrangement is access to AMD’s ROCm open source software environment, which will be made available within Vultr’s composable cloud infrastructure. This AMD-Vultr combo gives users access to thousands of open source, pre-trained AI models & frameworks.

Rockin’ with ROCm

AMD’s latest update to the software is ROCm 7, introduced in July and now live and ready to use. Version 7 offers advancements that include big performance gains, advanced features for scaling AI, and enterprise-ready AI tools.

One big benefit of AMD ROCm is that its open software ecosystem eliminates vendor lock-in. And when integrated with Vultr, ROCm supports AI frameworks that include PyTorch and TensorFlow, enabling flexible, rapid innovation. Further, ROCm future-proofs AI solutions by ensuring compatibility across hardware, promoting adaptability and scalability.

AMD’s roadmap is another attraction for Vultr. AMD products on tap for 2026 include the Instinct 400 family (codename Helios), new EPYC CPUs (Venice) and an 800-Gbit NIC (Vulcano).

Conversely, Vultr is a big business for AMD. Late last year, a tech blog reported that Vultr’s first shipment of AMD Instinct MI300X GPUs numbered “in the thousands.”

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Related Content

Tech Explainer: What is agentic AI?

Featured content

Tech Explainer: What is agentic AI?

Find out how new artificial intelligence systems can make decisions and take actions autonomously—that is, without human intervention.

Learn More about this topic
  • Applications:
  • Featured Technologies:

We’re on the precipice of a major AI evolution. Welcome to the era of agentic AI.

The official definition of agentic AI is artificial intelligence capable of making autonomous decisions. That is, without human oversight or intervention.

You can imagine agentic AI as a robot on a mission. This robot has been designed to think like a human. Give it a goal, and the robot can then evaluate the ongoing situation, reacting intelligently in pursuit of that defined goal.

For example, imagine you’re planning a visit to wineries in California’s Napa Valley. A standard AI chatbot like ChatGPT could help you find the closest airport with car-rental agencies, identify which airlines fly there, and locate nearby hotels. But it would still be up to you to compare prices and actually make the reservations.

But what if instead, your robot could autonomously plan—and book!—the entire trip based on your preferences? For example, you might engage an agentic AI like AutoGPT by telling it something like this:

“I want to go to Napa Valley and visit wineries. I don’t want to spend more than $3,000. I prefer Chardonnay and Syrah wines. I once had a bad experience with American Airlines. It would be fun to drive a convertible. A 3-star hotel is fine as long as it’s got good reviews.”

The promise of agentic AI is that it would use that information to plan and book your trip. The agentic AI would find you the best flight, car and hotel by interacting with each company’s APIs or even their own agentic AI—here referred to as “other agents.” This is also known as machine-to-machine (M2M) communications.

Your robot agent could also make your reservations at vineyards with critically acclaimed Chardonnay and Syrah wines. And it might even plan your route using details as granular as the range of the discounted rag-top Ford Mustang it found near the airport.

Agentic AI for Organizations

This personal Napa Valley scenario is one of those nice-to-have kinds of things. But for organizations, agentic AI has far more potential. This technology could eventually transform every major industry and vertical market.

For example, a retailer might use agentic AI to autonomously adjust a product’s price based on the current inventory level, availability and competitive brands.

A manufacturer could use an AI agent to manage procurement and create dynamic forecasting, saving the company time and money.

And in the public sector, agentic AI could help a government agency better respond to public-health emergencies like the next global pandemic. The AI could model viral transmission patterns, then send additional resources to the areas that need them the most.

In each case, we’re talking about the potential for a tireless virtual robot workforce. Once you give an agentic AI a mission, it can proceed without any further human intervention, saving you countless hours and dollars.

Training: Standard AI vs. Agentic

For all types of AI, one big issue is training. That’s because an AI system on its own doesn’t really know anything. To be useful, it first has to be trained.

And with training, there’s a huge difference between the way you train a standard AI and the way you train an AI that’s agentic. It’s as dramatic as the difference between programming a calculator and onboarding a new (human) intern.

With a standard AI chatbot, the system is trained to answer questions based on a relatively narrow set of parameters. To accomplish this, engineers provide massive amounts of data via large language models (LLMs). They then train the bot through supervised learning. Eventually, inferencing enables the AI to make predictions based on user input and available data.

By contrast, training an agentic AI focuses on memory, autonomy, planning and using available tools. Here, LLMs are paired with prompt engineering, long-term memory systems and feedback loops. These elements work together to create a type of intelligent thought process—the kind you hope your new intern is capable of!

Then, at the inferencing stage, the AI does far more than just answer questions. Instead, agentic AI inferencing enables the system to interpret goals, create plans, ask for help and, ultimately, execute tasks autonomously.

Nuts and Bolts

The IT infrastructure that powers agentic AI is no different from the horsepower behind your average chatbot. There’s just a lot more of it.

That’s because agentic AI, in comparison with standard AI, makes more inference calls, reads and writes more files, and queries more APIs. It also engages a persistent memory. That way, the AI can continuously access collected information as it works towards its goals.

However, having a slew of GPUs and endless solid-state storage won’t be enough to sustain what will likely be the meteoric growth of this cutting-edge technology. As agentic AI becomes more vital, IT managers will need a way to feed the fast-growing beast.

Supermicro’s current H14 systems—they include the GPU A+ Server—are powered by AMD EPYC 9005-series processors and fitted with up to 8 AMD Instinct MI325X Accelerators. Supermicro has designed these high-performance solutions to tackle the most challenging AI workloads.

Looking ahead, at AMD’s recent “Advancing AI” event, CEO Lisa Su introduced Helios, AMD’s vision for agentic AI infrastructure. Su said Helios will deliver the compute density, memory bandwidth, performance and scale-out bandwidth needed for the most demanding AI workloads. What’s more, Helios will come packaged as a ready-to-deploy AI rack solution that accelerates users’ time to market.

Helios, planned for release in 2026, will use several forthcoming products: AMD Instinct MI400 GPUs, AMD 6th Gen EPYC CPUs, and AMD Pensando “Vulcano” network interface cards (NICs). All will be integrated in an OCP-compliant rack that supports both UALink and Ultra Ethernet. And eventually, Helios will appear in turnkey systems such as the Supermicro H14 series.

What’s Next?

What else does agentic AI have in store for us? While no one has a crystal ball, it’s reasonable to assume we’ll see increasingly sophisticated agents infiltrating nearly every aspect of our lives.

For instance, agentic AI could eventually develop the ability to work autonomously on long-term, multifaceted projects—everything from advertising campaigns to biomedical research.

Agentic AI is also likely to learn how to debug its own logic and develop new tools. These capabilities are referred to by the pros as self-reflection and self-improvement, respectively.

One day in the not-too-distant future, we could even see massive teams of specialized AI agents working together under a single robotic project manager.

Think this is starting to sound like “The Matrix”? You ain’t seen nothin’ yet.

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Related Content

Deploy GenAI with confidence: Validated Server Designs from Supermicro and AMD

Featured content

Deploy GenAI with confidence: Validated Server Designs from Supermicro and AMD

Learn about the new Validated Design for AI clusters from Supermicro and AMD. It can save you time, reduce complexity and improve your ROI.

Learn More about this topic
  • Applications:
  • Featured Technologies:

The task of designing, building and connecting a server system that can run today’s artificial intelligence workloads is daunting.

Mainly, because there are a lot of moving parts. Assembling and connecting them all correctly is not only complicated, but also time-consuming.

Supermicro and AMD are here to help. They’ve recently co-published a Verified Design document that explains how to build an AI cluster. The PDF also tells you how you can acquire an AMD-powered Supermicro cluster for AI pre-built, with all elements connected, configured and burned in before shipping.

Full-Stack for GenAI

Supermicro and AMD are offering a fully validated, full-stack solution for today’s Generative AI workloads. The system’s scale can be easily adjusted from as few as 16 nodes to as many as 1,024—and points in between.

This Supermicro solution is based on three AMD elements: the AMD Instinct MI325X GPU, AMD Pensando Pollara 400 AI network interface card (NIC), and AMD EPYC CPU.

These three AMD parts are all integrated with Supermicro’s optimized servers. That includes network cabling and switching.

The new Validated Design document is designed to help potential buyers understand the joint AMD-Supermicro solution’s key elements. To shorten your implementation time, the document also provides an organized plan from start to finish.

Under the Cover

This comprehensive report—22 pages plus a lengthy appendix—goes into a lot of technical detail. That includes the traffic characteristics of AI training, impact of large “elephant” flows on the network fabric, and dynamic load balancing. Here’s a summary:

  • Foundations of AI Fabrics: Remote Direct Memory Access (RDMA), PCIe switching, Ethernet, IP and Border Gateway Protocol (BGP).
  • Validated Design Equipment and Configuration: Server options that optimize RDMA traffic with minimal distance, latency and silicon between the RDMA-capable NIC (RNIC) and accelerator.
  • Scaling Out the Accelerators with an Optimized Ethernet Fabric: Components and configurations including the AMD Pensando Pollara 400 Ethernet NIC and Supermicro’s own SSE-T8196 Ethernet switch.
  • Design of the Scale Unit—Scaling Out the Cluster: Designs are included for both air-cooled and liquid-cooled setups.
  • Resource Management and Adding Locality into Work Placement: Covering the Simple Linux Utility for Resource Management (SLURM) and topology optimization including the concept of rails.
  • Supermicro Validated AMD Instinct MI325 Design: Shows how you can scale the validated design all the way to 8,000 AMD MI325X GPUs in a cluster.
  • Storage Network Validated Design: Multiple alternatives are offered.
  • Importance of Automation: Human errors are, well, human. Automation can help with tasks including the production of detailed architectural drawings, output of cabling maps, and management of device firmware.
  • How to Minimize Deployment Time: Supermicro’s Rack Scale Solution Stack offers a fully integrated, end-to-end solution. And by offering a system that’s pre-validated, this also eases the complexity of multi-vendor integration.

Total Rack Solution

Looking to minimize implementation times? Supermicro offers a total rack scale solution that’s fully integrated and end-to-end.

This frees the user from having to integrate and validate a multi-vendor solution. Basically, Supermicro does it for you.

By leveraging industry-leading energy efficiency, liquid and air-cooled designs, and global logistics capabilities, Supermicro delivers a cost-effective and future-proof solution designed to meet the most demanding IT requirements.

The benefits to the customer include reduced operational overhead, a single point of accountability, streamlined procurement and deployment, and maximum return on investment.

For onsite deployment, Supermicro provides a turnkey, fully optimized rack solution that is ready to run. This helps organizations maximize efficiency, lower costs and ensure long-term reliability. It includes a dedicated on-site project manager.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Related Content

Meet Supermicro’s newest AI servers, powered by AMD Instinct MI350 Series GPUs

Featured content

Meet Supermicro’s newest AI servers, powered by AMD Instinct MI350 Series GPUs

Supermicro’s new AI servers are powered by a combination of AMD EPYC CPUs and AMD Instinct GPUs.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Supermicro didn’t waste any time supporting AMD’s new Instinct MI350 Series GPUs. The same day AMD formally introduced the new GPUs, Supermicro announced two rack-mount servers that support them.

The new servers, members of Supermicro’s H14 generation of GPU optimized solutions, feature dual AMD EPYC 9005 CPUs along with the AMD Instinct MI350 series GPUs. They’re aimed at organizations looking to achieve a formerly tough combination: maximum performance at scale in their AI-driven data centers, but also a lower total cost of ownership (TCO).

To make the new servers easy to upgrade and scale, Supermicro has designed the new servers around its proven building-block architecture.

Here’s a quick look at the two new Supermicro servers:

4U liquid-cooled system with AMD Instinct MI355X GPU

This system, model number AS -4126GS-NMR-LCC, comes with a choice of dual AMD EPYC 9005 or 9004 Series CPUs, both with liquid cooling.

On the GPU front, users also have a choice of the AMD Instinct MI325X or brand-new AMD Instinct MI355X. Either way, this server can handle up to 8 GPUs.

Liquid cooling is provided by a single direct-to-chip cold plate. Further cooling comes from 5 heavy-duty fans and an air shroud.

8U air-cooled system with AMD Instinct MI350X GPU

This system, model number AS -8126GS-TNMR, comes with a choice of dual AMD EPYC 9005 or 9004 Series CPUs, both with air cooling.

This system also supports both the AMD Instinct MI325X and AMD Instinct MI350X GPUs. Also like the 4U server, this system supports up to 8 GPUs.

Air cooling is provided by 10 heavy-duty fans and an air shroud.

The two systems also share some features in common. These include PCIe 5.0 connectivity, large memory capacities (up to 2.3TB), and support for both AMD’s ROCm open-source software and AMD Infinity Fabric Link connections for GPUs.

“Supermicro continues to lead the industry with the most experience in delivering high-performance systems designed for AI and HPC applications,” says Charles Liang, president and CEO of Supermicro. “The addition of the new AMD Instinct MI350 series GPUs to our GPU server lineup strengthens and expands our industry-leading AI solutions and gives customers greater choice and better performance as they design and build the next generation of data centers.”

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Related Content

AMD presents its vision for the AI future: open, collaborative, for everyone

Featured content

AMD presents its vision for the AI future: open, collaborative, for everyone

Check out the highlights of AMD’s Advancing AI event—including new GPUs, software and developer resources.

Learn More about this topic
  • Applications:
  • Featured Technologies:

AMD advanced its AI vision at the “Advancing AI” event on June 12. The event, held live in the Silicon Valley city of San Jose, Calif., as well as online, featured presentations by top AMD executives and partners.

As many of the speakers made clear, AMD’s vision for AI is that it be open, developer-friendly, collaborative and useful to all.

AMD certainly believes the market opportunity is huge. During the day’s keynote, CEO Lisa Su said AMD now believes the total addressable market (TAM) for data-center AI will exceed $500 billion by as soon as 2028.

And that’s not all. Su also said she expects AI to move beyond the data center, finding new uses in edge computers, PCs, smartphone and other devices.

To deliver on this vision, Su explained, AMD is taking a three-pronged approach to AI:

  • Offer a broad portfolio of compute solutions.
  • Invest in an open development ecosystem.
  • Deliver full-stack solutions via investments and acquisitions.

The event, lasting over two hours, was also filled with announcements. Here are the highlights.

New: AMD Instinct MI350 Series

At the Advancing AI event, CEO Su formally announced the company’s AMD Instinct MI350 Series GPUs.

There are two models, the MI350X and MI355X. Though both are based on the same silicon, the MI355X supports higher thermals.

These GPUs, Su explained, are based on AMD’s 4th gen Instinct architecture, and each GPU comprises 10 chiplets containing a total of 185 billion transistors. The new Instinct solutions can be used for both AI training and AI inference, and they can also be configured in either liquid- or air-cooled systems.

Su said the MI355X delivers a massive 35x general increase in AI performance over the previous-generation Instinct MI300. For AI training, the Instinct MI355X offers up to 3x more throughput than the Instinct MI300. And in comparison with a leading competitive GPU, the new AMD GPU can create up to 40% more tokens per dollar.

AMD’s event also featured several representatives of companies already using AMD Instinct MI300 GPUs. They included Microsoft, Meta and Oracle.

Introducing ROCm 7 and AMD Developer Cloud

Vamsi Boppana, AMD’s senior VP of AI, announced ROCm 7, the latest version of AMD’s open-source AI software stack. ROCm 7 features improved support for industry-standard frameworks; expanded hardware compatibility; and new development tools, drivers, APIs and libraries to accelerate AI development and deployment.

Earlier in the day, CEO Su said AMD’s software efforts “are all about the developer experience.” To that end, Boppana introduced the AMD Developer Cloud, a new service designed for rapid, high-performance AI development.

He also said AMD is giving developers a 25-hour credit on the Developer Cloud with “no strings.” The new AMD Developer Cloud is generally available now.

Road Map: Instinct MI400, Helios rack, Venice CPU, Vulcano NIC

During the last segment of the AMD event, Su gave attendees a sneak peek at several forthcoming products:

  • Instinct MI400 Series: This GPU is being designed for both large-scale AI inference and training. It will be the heart of the Helios rack solution (see below) and provide what Su described as “the engine for the next generation of AI.” Expect performance of up to 40 petaflops, 432GB of HBM4 memory, and bandwidth of 19.6TB/sec.
  • Helios: The code name for a unified AI rack solution coming in 2026. As Su explained it, Helios will be a rack configuration that functions like a single AI engine, incorporating AMD’s EPYC CPU, Instinct GPU, Pensando Pollara network interface card (NIC) and ROCm software. Specs include up to 72 GPUs in a rack and 31TB of HBM3 memory.
  • Venice: This is the code name for the next generation of AMD EPYC server CPUs, Su said. They’ll be based on a 2nm form, feature up to 256 cores, and offer a 1.7x performance boost over the current generation.
  • Vulcano: A future NIC, it will be built using a 3nm form and feature speeds of up to 800Gb/sec.

Do More:

 

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Related Content

AI across AMD’s entire portfolio? Believe it!

Featured content

AI across AMD’s entire portfolio? Believe it!

A little over a year ago, AMD CTO Mark Papermaster said the company’s strategy was to offer AI everywhere. Now learn how AMD, with help from Supermicro, is bringing this strategy to life.

Learn More about this topic
  • Featured Technologies:

A year in the fast-moving world of artificial intelligence can seem like a lifetime.

Consider:

  • A year ago, ChatGPT had fewer than 200 million weekly active users. Now this Generative AI tool has 400 million weekly users, according to developer OpenAI.
  • A year ago, no one outside of China had heard of DeepSeek. Now its GenAI chatbot is disrupting the AI industry, challenging the way some mainstream tools function.
  • About a year ago, AMD CTO Mark Papermaster said his company’s new strategy called for AI across the entire product portfolio. Now AMD, with help from Supermicro, offers AI power for the data center, cloud and desktop. AMD also offers a robust open AI stack.

‘We’re Thrilled’

AMD’s Papermaster made his comments in Feb. 2024 during a fireside chat hosted by stock research firm Arete Research.

During the interview, CTO Papermaster acknowledged that most early customers for AMD’s AI hardware were mostly big cloud hyperscalers, including AWS, Google Cloud and Microsoft Azure. But he also said new customers are coming, including both enterprises and individual endpoint users.

“We’re thrilled to bring AI across our entire portfolio,” Papermaster said.                                                                          

So how has AMD done? According to the company’s financial results for both the fourth quarter and the full year 2024, pretty good.

Aggressive Investments

During AMD’s recent report on its Q4:24 and full-year ’24 financial results, CFO Jean Hu mentioned that the company is “investing aggressively in AI.” She wasn’t kidding, as the following items show:

  • AMD is accelerating its AI software road map. The company released ROCm 6.3, which includes enhancements for faster AI inferencing on AMD Instinct GPUs. The company also shared an update on its plans for the ROCm software stack.
  • AMD announced a new GPU system in 2024, the AMD Instinct MI325X. Designed for GenAI performance, it’s built on the AMD CDNA3 architecture and offers up to 256GB of HBM3E memory and up to 6TB/sec. of bandwidth.
  • To provide a scalable AI infrastructure, AMD has expanded its partnerships. These partnerships involve companies that include Aleph, IBM, Fujitsu and Vultr. IBM, for one, plans to deploy AMD MI300X GPUs to power GenAI and HPC applications on its cloud offering.
  • AMD is offering AI power for PCs. The company added AI capabilities to its Ryzen line of processors. Dell, among other PC vendors, has agreed to use these AMD CPUs in its Dell Pro notebook and desktop systems.

Supermicro Servers

AMD partner Supermicro is on the AI case, too. The company now offers several AMD-powered servers designed specifically for HPC and AI workloads.

These include an 8U 8-GPU system with AMD Instinct MI300X GPUs. It’s designed to handle some of the largest AI and GenAI models.

There’s also a Supermicro liquid-cooled 2U 4-way server. This system is powered by the AMD Instinct MI300A, which combines CPUs and GPUs, and it’s designed to support workloads that coverge HPC and AI.

Put it all together, and you can see how AMD is implementing AI across its entire portfolio.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Related Content