Sponsored by:

Visit AMD Visit Supermicro

Performance Intensive Computing

Capture the full potential of IT

Tech Explainer: What are CPU Cores, Threads, Cache & Nodes?

Featured content

Tech Explainer: What are CPU Cores, Threads, Cache & Nodes?

Today’s CPUs are complex. Find out what the key components actually do—and why, in an age of AI, they still matter.

Learn More about this topic
  • Applications:
  • Featured Technologies:

In the age of artificial intelligence, CPUs still matter. A central processor’s parts—cores, threads, cache and nodes—are as important as any AI accelerator.

But what exactly do those CPU parts do? And why, in an age of AI, do they still matter?

These questions are easy to overlook given AI’s focus on the GPU. To be sure, graphical processors are important for today’s AI workloads. But the humble CPU also remains a vital component.

If the GPU is AI’s turbocharger, then the CPU is the engine that makes the whole car go. As Dan McNamara, AMD’s GM of compute and enterprise AI business, said at the recent AMD Financial Analysts Day, “AI requires leadership CPUs.”

So here’s a look at the most important components of today’s data-center x86 CPUs. And an explanation of why they matter.

Cores: Heavy Lifting

The central processing unit is the brain of any PC or server. It reads instructions, does the complex math, and coordinates the system’s every task.

Zooming into the architecture of a CPU, it’s the individual cores that put the “PU” in CPU. Each fully independent processing unit can run its own task, virtual machine (VM) or container.

Modern enterprise-class CPUs such as AMD’s EPYC 9005 Series offer anywhere from 8 to 192 cores each. They operate at up to 5GHz.

These cores are built using AMD’s ‘Zen’ architecture. It’s a fundamental core design that offers enhancements vital to data centers, including improved instructions-per-clock (IPC), branch prediction, caches and efficiency.

Performance like that is a must-have when it comes to a data center’s most demanding tasks. That’s especially true for compute-intensive database operations and API-heavy microservices such as authentication, payment gateways and search.

Having more cores in each CPU also enables IT managers to run more workloads per server. That, in turn, helps organizations lower their hardware and operating costs, simplify IT operations, and more easily scale operations.

Threads: Helping Cores Do More

A modern CPU core needs to multitask, and that’s where having multiple threads is essential. A single CPU core with two threads can juggle two tasks by switching between them very quickly. In a CPU with a high core count, a productivity-multiplier like that becomes exponentially more effective.

This capability delivers two important benefits. One, it helps ensure that each CPU core stays productive, even if one task stalls. And two, it boosts the CPU’s overall output.

For example, the AMD EPYC 9965 processor boasts 192 cores with a total of 384 threads. That kind of multitasking horsepower helps smooth request handling for web services and microservices. It also improves VM responsiveness and helps AI workloads run more efficiently under heavy loads.

Cache: Speedy but Short-Term Memory

The unsung heroes of CPU design? That would be cache.

The main job of a CPU cache is to help the cores juggle data with low latency. Remember, less latency is always better.

As a result, CPU cache enables databases to run faster, improve VM density and reduce latency.

Your average CPU cache is arranged in three layers:

  • L1 cache is very small and very fast. Each core has its own L1 cache, which holds around 32 KB of instructions and data. The L1 cache sends that data to a register— a tiny, ultra-fast storage location the core uses to acquire the data used for calculations.
  • L2 cache is also exclusive to each core. At around 1MB, this cache is bigger than L1, but it’s also a little slower. L2 cache holds any data that doesn’t fit in the L1 cache. Working together, the L1 and L2 caches can quickly pass data back and forth until ultimately, the L1 cache passes the data to the core.
  • L3 cache is shared by all cores in a CPU, and it acts as a buffer for passing data between the CPU and main memory. Sizes vary widely. In an 8-core AMD EPYC processor, the L3 cache is just 64MB. But in AMD’s 192-core CPU, the L3 Cache gets as big as 348MB.

Some AMD CPUs, including the AMD EPYC 9845, also include a 3D V-Cache. This AMD innovation stacks an additional cache on top of the L3 cache (hence the name 3D). Stacking the two caches vertically adds storage without increasing the overall size of the CPU.

The added 3D V-Cache also improves performance for workloads that benefit from a larger cache. Examples include scientific simulations and big data.

Nodes: Power & Efficiency

When it comes to CPU nodes, smaller is better. A smaller node size can deliver benefits that include lower power consumption, increased efficiency, and more compute performance per watt.

Nodes are expressed in nanometers (nm)—that’s one billionth of a meter—which describe the tiny size of transistors on a chip.

The latest AMD EPYC 9005-series architectures, ‘Zen 5’ and ‘Zen 5c,’ are built on 4nm and 3nm nodes, respectively.

Each of these individual performance gains may seem tiny when considered on a per-chip basis. But in the aggregate, they can make a huge difference. That’s especially true for resource-intensive workloads such as AI training and inferencing.

Coming Soon: Smaller, Faster CPUs

AMD’s near-term roadmap tells us we can expect its AMD EPYC CPUs to continue getting smaller, faster and more efficient.

Those manufacturing and performance gains will likely come from more cores per CPU socket, bigger and more efficient caches. Earlier this year, AMD said the next generation of its EPYC processors, codenamed Venice, will be brought up on TSMC’s advanced 2nm process technology.

Enterprises will be able to parlay those improvements into better performance under multi-tenant loads and reduced latency overall. The latter is particularly vital for modern operations.

The bottom lie: Denser CPU cores mean big business, both for processor makers such as AMD and the server vendors such as Supermicro that rely on these CPUs.

Denser CPUs are also vital for enterprises now transforming their data centers for AI. Because adding space is so slow and costly, these organizations are instead looking to pack more compute power per rack. Smaller, more powerful CPUs are an important part of their solution.

Minimum CPU size with maximum power? It’s coming soon to a data center near you.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Check out Supermicro’s new AMD GPU-powered server—it’s air-cooled

Featured content

Check out Supermicro’s new AMD GPU-powered server—it’s air-cooled

Supermicro’s new 10U server is powered by AMD’s EPYC CPUs and Instinct MI355X GPUs. And it’s kept cool by nearly 20 fans.

Learn More about this topic
  • Applications:
  • Featured Technologies:

What do you do if you need GPU power for AI and other compute-intensive workloads, but lack the infrastructure for liquid cooling?

Supermicro has the answer. The company just introduced a 10U server powered by AMD Instinct MI355X GPUs that’s air-cooled.

The new server, showcased at the recent SC25 conference in St. Louis, is Supermicro model AS -A126GS-TNMR.

Each server is powered by the customer’s choice of dual AMD EPYC 9004 or 9005 Series CPUs with up to 384 cores and 768 threads. The system also features a total of eight AMD Instinct MI355X onboard OAM GPU accelerator modules, which are air-cooled. (OAM is short for OCP Accelerator Module, an industry-standard form factor for AI hardware.) In addition, these accelerated GPU servers offer up to 6TB of DDR5 system memory.

While the systems are air-cooled with up to 19 heavy-duty fans, there’s no penalty in terms of cooling capacity. In fact, AMD has boosted the GPU’s thermal design point (TDP)—the maximum amount of heat a server’s cooling system can handle—from 1000W to 1400W.

Also, compared with the company’s air-cooled 8U server based on AMD Instinct MI350X GPUs, the 10U server offers up to double-digit more performance, according to Supermicro . For end users, that means faster data processing.

More Per Rack

The bigger picture: Supermicro’s new 10U option lets customers unlock higher performance per rack. And with their choice of 10U air cooling or 4U liquid cooling, both powered by the latest AMD EPYC processors.

Supermicro’s GPU solutions are designed to offer maximum performance for AI and inference at scale. And they’re intended for use by both cloud service providers and enterprises.

Are your customers looking for a GPU-powered server that’s air cooled? Tell them about these new Supermicro 10U servers. And let them know that these systems are ready to ship now.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Tech Explainer: What’s new in AMD ROCm 7?

Featured content

Tech Explainer: What’s new in AMD ROCm 7?

Learn how the AMD ROCm software stack has been updated for the era of AI.

Learn More about this topic
  • Applications:
  • Featured Technologies:

While GPUs have become the digital engines of our increasingly AI-powered lives, controlling them accurately and efficiently can be tricky.

That’s why, in 2016, AMD created ROCm. Pronounced rock-em, it’s a software stack designed to translate the code written by programmers into sets of instructions that AMD GPUs can understand and execute perfectly.

If the GPUs in today’s cutting-edge AI servers are the orchestra, then ROCm is the sheet music being played.

AMD introduced the latest version, ROCm 7.0, earlier this fall. Version 7.0 is designed for the new world of AI.

How ROCm works

ROCm is a platform created by AMD to run programs on its AI-focused GPUs, the Instinct MI350 Series accelerators. AMD calls the latest version, ROCm 7.0, an AI-ready powerhouse designed for performance, efficiency and productivity.

Providing that kind of facility is a matter of far more than just simple software. ROCm is actually an expansive collection of tools, drivers and libraries.

What’s in the collection? The full ROCm stack contains:

  • Drivers that enable a computer’s operating system to communicate with any installed AMD GPUs.
  • The Heterogeneous Interface for Portability (HIP), a coding system for users to create and run custom GPU programs.
  • Math and AI libraries including specialized tools like deep learning operations, fast math routines, matrix multiplication, and tensor ops. These AI building blocks are pre-built to help developers accelerate production.
  • Compilers that turn code into GPU instructions.
  • System-management tools that developers can use to debug applications and optimize GPU performance.

Help Me, GPU

The latest version of ROCm is purpose-built for generative AI and large-scale AI inferencing and training. While developers rely on GPUs for parallel processing, performing many tasks at once, GPUs are general-purpose processors. To achieve the best performance for AI workloads, developers need a software bridge that turns their high-level code into GPU-optimized instructions. That bridge is ROCm.

ROCm lets developers run AI frameworks that include PyTorch effectively on AMD GPUs. ROCm converts application code into instructions designed for the hardware. In this way, ROCm helps organizations improve performance, scale workloads across multiple GPUs, and meet increasing demand without sacrificing reliability.
 
For demanding AI workloads such as those using Mixture of Experts (MoE) models, ROCm is essential for execution. MoE models activate only a small group of expert networks for each input, resulting in sparse workloads that are efficient, but hard to schedule. ROCm ensures that GPUs can perform these sparse operations at scale, maintaining high throughput and accuracy across clusters.
 
In other words, ROCm provides the tools and runtime to make even the most complex GPU workloads run efficiently. It connects AI developers with the hardware that supports their applications.
 
That’s important. While increased demand is what every enterprise wants, it still brings challenges that leave little room for mistakes.
 
Open Source Power

But wait, there's more. AMD ROCm has another clever trick up its sleeve: open-source integration.

By using popular open-source frameworks, ROCm lets enterprises and developers run large-scale inference workloads more efficiently. This open source approach also empowers the same organizations and developers to break free of proprietary software and vendor-locked ecosystems.

Free from those dependencies, users can scale AI clusters by deploying commodity components instead of being locked into a single vendor’s hardware. Ultimately, that can lead to lower hardware and licensing costs.

This approach also empowers users to customize their AI operations. In this way, AI systems can be developed to better suit the unique requirements of an organization’s applications, environments and end users.

Another Layer

While ROCm serves the larger market, the recent release of AMD’s new Enterprise AI Suite shows the company’s commitment to developing tools specifically for enterprise-class organizations.

AMD says the new suite can to take enterprises from bare metal server to enterprise-ready AI software in mere minutes.

To accomplish this, the suite provides four additional components: solution blueprints, inference microservices, AI Workbench, and a dedicated resource manager.

These tools are designed to help enterprises better scale their AI workloads, predict costs and capacity, and accelerate time-to-production.

Always Be Developing

Along with these product releases, AMD is being perfectly clear about its focus on AI development. At the company’s recent Financial Analyst Day, AMD CEO Lisa Su explained that over the last five years, the cost of AMD’s AI-related investments and acquisitions has topped $100 billion. That includes building up a staff of some 25,000 engineers.

Looking ahead, Su told financial analysts that AMD’s data-center AI business is on track to draw revenue in the “tens of billions of dollars” by 2027. She also said that over the next three to five years, AMD expects its data-center AI revenue to enjoy a compound annual growth rate (CAGR) of over 80%.

AMD’s roadmap points to updates that will focus on further boosts to performance, productivity and scalability. The company may accomplish these gains by offering more streamlined build and packaging systems, more optimized training and inferencing, and broader hardware support. It’s also reasonable to expect improved virtualization and multi-tenant support.

That said, if you want your speculation about future AI-centric ROCm improvements to be as accurate as possible, your best bet may be to ask an AI chatbot…powered by Supermicro and AMD, of course.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Supermicro adds MicroBlade for CSPs powered by AMD EPYC 4005 series processors

Featured content

Supermicro adds MicroBlade for CSPs powered by AMD EPYC 4005 series processors

To serve cloud service providers, Supermicro adds a 6U, 20-node MicroBlade server powered by AMD EPYC 4005 series processors.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Not every cloud service provider is as big or deep-pocketed as the big three—AWS, Google and Microsoft. And to serve those smaller and midsize CSPs, Supermicro recently added a 6U, 20-node server to its MicroBlade family powered by AMD EPYC 4005 series processors.

Smaller CSPs represent a big market. To be sure,  AWS, Google and Microsoft collectively drew nearly 65% of total worldwide cloud infrastructure services revenue in this year’s third quarter, according to Synergy Research Group.

But for both smaller CSPs and their suppliers, that remaining 35% was still quite valuable. Synergy estimates worldwide cloud infrastructure services revenue in Q3 totaled $107 billion. That means the share left to smaller and midsize CSPs was about $37 billion.

MicroBlade, Macro Benes

To serve these smaller CSPs, Supermicro recently introduced a 6U, 20-node MicroBlade (model number MBA-315R-1G) powered by a single AMD EPYC 4005 series processor.

This MicroBlade system delivers a cost-effective, green computing solution. It’s intended for workloads that include not only cloud computing, but also web hosting, dedicated hosting, virtual desktop infrastructure (VDI), AI inferencing, and enterprise workloads.

Supermicro CEO Charles Liang calls the new servers “a very cost-effective, green computing solution for cloud service providers.”

Key benefits of the new Supermicro system include up to 95% cable reduction with two integrated Ethernet switches per server; 70% space savings; and 30% energy savings over traditional 1U servers.

The system offers 3.3x higher density than a traditional 1U server. As a result, users can pack as many as 160 servers with 2,650 CPU cores, as well as 16 Ethernet switches, in a single 48U rack.

Under the hood, each MicroBlade server blade supports a single AMD EPYC 4005 CPU with up to 16 cores and 192GB of DDR5 memory. Also supported is a dual-slot, full-height/full-length (FHFL) GPU.

Also, this Supermicro system contains a dual-port 10GbE network switch. It’s designed to simplify topologies and enable more server instances per rack.

The 6U MicroBlade chassis can hold up to 20 individual server blades, two Ethernet switches and two management modules.

To protect workloads such as dedicated hosting, VDI, online gaming and AI inferencing, the Supermicro system also offers N+N redundancy. This setup configures two sets of independent components to provide high levels of reliability.

The MicroBlade system will also be available as a motherboard (model number BH4SRG) for Supermicro A+ servers.

 

 

Inside the AMD EPYC 4005 Series

The processors powering the new Supermicro server, AMD’s EPYC 4005 series, offer powerful performance for AI, cloud and hosting workloads. Yet they’re attractively priced for smaller businesses and hosting services.

The processors are based on the same core generation, ‘Zen 5,’ as are AMD’s more powerful data center processors, the AMD EPYC 9005 series. Yet the 4005 series processors have been designed for smaller operations, offering a combination of affordability, efficiency and ease of use.

AMD’s corporate VP for enterprise and HPC, Derek Dicker, says the AMD EPYC 4005 series processors “give our technology partners the flexibility to create powerful yet affordable systems that meet the specific needs of growing businesses and dedicated hosters.”

Do you have CSP clients looking for an affordable yet powerful servers? Tell them about these new AMD-powered Supermicro servers, coming soon.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Retail AI at the edge: Now here from Supermicro, AMD & Wobot.ai

Featured content

Retail AI at the edge: Now here from Supermicro, AMD & Wobot.ai

Retailers can now use AI to analyze in-store videos, thanks to a new system from Supermicro, AMD and Wobot.ai.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Artificial intelligence is being adapted for specific industry verticals. That now includes retail.

Supermicro, AMD and Wobot Intelligence Inc., a video intelligence supplier, are partnering to provide retailers with a short-depth server they can use to drive AI-powered analysis of their in-store videos. With these analyses, retailers can improve store operations, elevate the customer experience and boost sales.

The new server system was recently showcased by the three partners at NRF Europe 2025, an international conference for retailers. This year’s NRF Europe was held in Paris, France, in mid-September.

The new retail system is based on a Supermicro 1U server, model AS -1115S-FWTRT. It’s a short-depth front I/O system powered by a single AMD EPYC 8004 processor.

The server’s other features include dual 10G ports, dual 2.5-inch drive bays, up to 768GB of DDR5 memory, and an 800W redundant platinum power supply. This server is air-cooled by as many as six heavy-duty fans, and it supports a pair of single-width GPUs.

Good to Go

The retail system’s video-analysis software, provided by Wobot.ai, features a single dashboard, performance benchmarking, and easy installation and configuration. It’s designed to work with a user’s existing CCTV setup.

The company’s WoConnect app helps users connect digital video recorders (DVRs) and network video recorders (NVRs) in their private network to their Wobot.ai account. The app routes the user’s camera feeds to the AI.

Target use cases for retailers include store operations, loss prevention and compliance, customer behavior and footfall analysis.

More specifically, retailers can use the system to conduct video analyses that include:

  • Zone-based analytics: Which areas of the store draw the most attention? Which products draw interaction? How do customers move through the store?
  • Heat maps and event tracking: Visualize “crowd magnets” to improve future sales.
  • Customer-path analysis: Observe which sections of the store customers explore the most, and also see where they linger.

Using the system, retailers can enjoy a long list of benefits that include accelerated checkout processes, fewer customer walkaways, fine-tuned staffing levels, and improved product placement.

For example, a chain of juice bars with nearly 145 locations in California turned to Wobot.ai for help speeding customer service and improving employee productivity. Based on its video analyses, the retailer worked with Wobot.ai to design a pilot program for 10 stores. In just three months, the pilot delivered additional revenue in the test stores equivalent to 2% to 2.5% a year.

Wobot.ai also offers its video intelligence systems to other verticals, including hospitality, food service and security.

Edgy

One important feature of the new server is that it allows retailers to run real-time AI-powered video analysis at the edge. The Supermicro server is housed in a short-depth form factor, meaning it can be run in retail sites that lack a full-fledged data center.

Similarly, the system’s AMD EPYC 8004 processor has been optimized for power efficiency—important for installations at the edge. Featuring up to 64 ‘Zen4c’ dense cores, this AMD processor is specifically designed for intelligent edge and communications workloads.

By processing the AI analysis on-premises, the new system also offers low latency and high levels of privacy. Wobot.ai says its software is scalable across literally thousands of locations.

And the software is designed to be integrated easily with retailers’ existing camera infrastructure. In this way, it offers fast time-to-value and a quick return on investment.

Do you have retail customers looking for an edge—with AI at the edge? Tell them about this new retail solution today.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

4 IT events this fall you won’t want to miss

Featured content

4 IT events this fall you won’t want to miss

Important IT industry events are coming in October and November--with lots of participation from AMD and Supermicro. 

Learn More about this topic
  • Applications:
  • Featured Technologies:

Summer’s over…somehow it’s already October…and that means it’s time to attend important IT industry conferences, summits and other get-togethers.

Here’s your Performance Intensive Computing preview of four top events coming this month and next.

OCP Global Summit

  • Where & when: San Jose, California; Oct. 13-16, 2025
  • Who it’s for: This event, sponsored by the Open Compute Project (OCP), is for anyone interested in redesigning open source hardware to support the changing demands on compute infrastructure. This year’s theme: “Leading the future of AI.”
  • Who will be there: Speakers this year include Vik Malyala, senior VP of technology and AI at Supermicro; Mark Papermaster, CTO of AMD; Johnson Eung, staff growth product manager in AI at Supermicro; Shane Corban, senior director of technical product management at AMD; and Morris Ruan, director of product management at Supermicro.
     
  • Fun facts: AMD is a Diamond sponsor, and Supermicro is an Emerald sponsor.

~~~~~~~~~~~~~~~~~~~~

AMD AI Developer Day

  • Where & when: San Francisco, Oct. 20, 2025
  • Who it’s for: Developers of artificial intelligence applications and systems. Workshop topics will include developing multi-model, multi-agent systems; generating videos using open source tools; and developing optimized kernals.
  • Who will be there: Speakers will include executives from the University of California, Berkeley; Red Hat AI; Google DeepMind; and OpenAI. Also speaking will be execs from Ollama, an open source platform for AI models; Unsloth AI, an open source AI startup; vLLM, a library for large language model (LLM) inference and serving; and SGLang, an LLM framework.
  • Fun facts:
    • Supermicro is a conference sponsor.
    • During the conference, winners of the AMD Developer Challenge will be announced. The grand prize winner will take home $100,000.
    • AMD, PyTorch and Unsloth AI are co-sponsoring a virtual hackathon, the Synthetic Data AI Agents Challenge, on Oct. 18-20. The first-prize winners will receive $3,000 plus 1,200 hours of GPU credits.

~~~~~~~~~~~~~~~~~~~~

AI Infra Summit

  • Where & when: San Francisco; Nov. 7, 2025
  • Who it’s for: Anyone interested in the convergence of AI innovation and scalable infrastructure. This event is being hosted by Ignite, a go-to-market provider for the technology industry.
  • Who will be there: The speaker lineup is still TBA, but is promised to include enterprise technology leaders, AI and machine learning engineers, cloud and data center architects, venture capital investors, and infrastructure vendors.
  • Fun facts:
    • This is a hybrid event. You can attend either live or online.
    • AMD and Supermicro are Stadium-level sponsors.

~~~~~~~~~~~~~~~~~~~~

SC25

  • Where & when: St. Louis, Missouri; Nov. 16-21, 2025
  • Who it’s for: The global supercomputing community, including those working in high performance computing (HPC), networking, storage and analysis. This year’s theme: “HPC ignites.”
  • Who will be there: Speakers will feature nearly a dozen AMD executives, including Rob Curtis, a Fellow in Data Center Platform Engineering; Shelby Lockhart, a software system engineer; and Nuwan Jayasena, a Fellow in AMD Research. They and other speakers will appear in panels, presentations of papers, workshops, tutorials and more.
     
  • Fun facts: SC25 will feature a series of noncommercial “Birds of a Feather” sessions that allow attendees to openly discuss topics of mutual interest.

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Vultr, Supermicro, AMD team to offer hi-performance cloud compute & AI infrastructure

Featured content

Vultr, Supermicro, AMD team to offer hi-performance cloud compute & AI infrastructure

Vultr, a global provider of cloud services, now offers Supermicro servers powered by AMD Instinct GPUs.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Supermicro servers powered by the latest AMD Instinct GPUs and supported by the AMD ROCm open software ecosystem are at the heart of a global cloud infrastructure program offered by Vultr.

Vultr calls itself a modern hyperscaler, meaning it provides cloud solutions for organizations facing complex AI and HPC workloads, high operational costs, vendor lock-in, and the need for rapid insights.

Launched in 2014, Vultr today offers services from 32 data centers worldwide, which it says can reach 90% of the world’s population in under 40 milliseconds. Vultr’s services include cloud instances, dedicated servers, cloud GPUs, and managed services for database, cloud storage and networking.

Vultr’s customers enjoy benefits that include costs 30% to 50% lower than those of the hyperscalers and 20% to 30% lower than those of other independent cloud providers. These customers—there are over 220,000 of them worldwide—also enjoy Vultr’s full native AI stack of compute, storage and networking.

Vultr is the flagship product of The Constant Co., based in West Palm Beach, Fla. The company was founded by David Aninowsky, an entrepreneur who also started GameServers.com and served as its CEO for 18 years.

Now Vultr counts among its partners AMD, which joined the Vultr Cloud Alliance, a partner program, just a year ago. In addition, AMD’s venture group co-led a funding round this past December that brought Vultr $333 million.

Expanded Data Center

Vultr is now expanding its relationship with Supermicro, in part because that company is first to market with the latest AMD Instinct GPUs. Vultr is now offering Supermcro systems powered by AMD Instinct MI355X, MI325X and MI300X GPUs. And as part of the partnership, Supermicro engineers work on-site with Vultr technicians.

Vultr is also relying on Supermicro for scaling. That’s a challenge for large AI implementations, as these configurations require deep expertise for both integration and operations.

Among Vultr’s offerings from Supermicro is a 4U liquid-cooled server (model AS -4126GS-NMR-LCC) with dual AMD EPYC 9005/9004 processors and up to eight AMD GPUs—the user’s choice of either MI325X or MI355X.

Another benefit of the new arrangement is access to AMD’s ROCm open source software environment, which will be made available within Vultr’s composable cloud infrastructure. This AMD-Vultr combo gives users access to thousands of open source, pre-trained AI models & frameworks.

Rockin’ with ROCm

AMD’s latest update to the software is ROCm 7, introduced in July and now live and ready to use. Version 7 offers advancements that include big performance gains, advanced features for scaling AI, and enterprise-ready AI tools.

One big benefit of AMD ROCm is that its open software ecosystem eliminates vendor lock-in. And when integrated with Vultr, ROCm supports AI frameworks that include PyTorch and TensorFlow, enabling flexible, rapid innovation. Further, ROCm future-proofs AI solutions by ensuring compatibility across hardware, promoting adaptability and scalability.

AMD’s roadmap is another attraction for Vultr. AMD products on tap for 2026 include the Instinct 400 family (codename Helios), new EPYC CPUs (Venice) and an 800-Gbit NIC (Vulcano).

Conversely, Vultr is a big business for AMD. Late last year, a tech blog reported that Vultr’s first shipment of AMD Instinct MI300X GPUs numbered “in the thousands.”

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Retail in the Spotlight: Making Shelf Space for AI

Featured content

Retail in the Spotlight: Making Shelf Space for AI

Learn how retailers including Amazon, Sephora and Walmart are applying artificial intelligence to deliver real business benefits—and help their shoppers find just the right product.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Retailers are relying more and more on artificial intelligence. And the reason is simple: AI technology can help retailers engage customers, lower operational costs and increase revenue.

Indeed, over 70% of retailers anticipate a significant ROI from AI in the next year, according to accounting firm KPMG.

Their customers approve of AI, too. In a poll conducted earlier this year by vision AI provider Everseen, two out of three consumers said AI makes shopping more convenient.

That’s a true win-win scenario.

Customer-facing

On the retail customer side, AI provides helpful features such as support chatbots and personal shopping assistants. AI can also offer visual search, letting customers upload photos of products they like and find similar items in real time.

AI is also capable of creating personalized recommendations that go far beyond the typical “people who bought X also bought Y” message.

For example, the AI behind Amazon’s industry-leading recommendation engine takes into account a customer’s shopping habits all the way back to the time they first created an account. Then the engine combines that data with whatever demographic information it can dig up or infer. The result: Customers receive genuinely useful suggestions.

Amazon also has a retail-focused chatbot called Rufus that can answer online shoppers’ questions about a products they haven’t bought yet, but are only considering. To do this work, the GenAI-powered shopping assistant has been trained on a potent mix of data that includes the entire Amazon catalog, customer reviews, community interviews and information from the public web.

This lets consumers ask Rufus just about anything. For example, “Are these shoes good for narrow feet?” will get an answer. And so will “Can this sharpener create the 16-degree angle recommended by the maker of my fancy Japanese chef’s knife?”

If you’re looking for a bit more wow factor, consider the Sephora Virtual Artist. This AI-powered virtual try-on feature uses your smartphone’s augmented reality (AR) to show how you’d look with a particular shade of lipstick, eye shadow or other makeup.

Don’t care for one shade? Sephora’s AI will suggest a better one based on your skin tone. Then it will find your color in stock at a store near you—along with a complimentary foundation, blush and eye liner.

Behind the Scenes

Deploying AI helps retailers save time and money. That’s especially true for those with big warehouses and complex supply chains.

Both Walmart and Amazon employ small armies of AI-enabled robots to zip around their warehouses. These tireless heavy-lifters find what they’re looking for by scanning bar- and QR-codes. Once they locate a product, their robotic arms grab it off even the highest shelf. Then the robots efficiently transport the products to their shipping departments.

These AI-powered robots can also report to other parts of the system, many of which use AI as well. One example is an inventory-control AI module that forecasts demand and makes sure the warehouse stays well-stocked. Another is a bot designed to manage complex supply chains by calculating trends, market prices, availability and shipping times.

Increasingly, retailers rely on AI for marketing too. They use retail bots to keep an eye on customer sentiment and emerging trends by scraping online reviews and social media posts. This information can also help retailers deal with customer-service issues before they get out of hand. And AI systems provide vital market data that retailers can use as they plan and launch new product lines.

Retail Power

Retail AI software is hugely powerful, but the hardware matters too. Deprived of enough power to collect, analyze and act on terabytes of daily data, AI is just reams of pointless code.

So retailers rely on purpose-built retail AI hardware solutions. That includes the Supermicro AS -2115HE-FTNR server.

This retail AI-server is powered by 5th gen AMD EPYC processors and has room for up to 6TB of ECC DDR5 memory and four GPUs. Retailers can also configure the system with up to 6 hot-swappable drives and their choice of air or liquid cooling.

The improved density in Supermicro’s multi-node racks helps retail organizations achieve a lower total cost of ownership by reducing server counts and energy demands.

Retail’s Future

AI is becoming more sophisticated every day. Soon, powerful new features will catalyze a paradigm shift in retail operations.

As agentic AI changes from a fascinating new design to a daily mainstay, hyper-personalized, frictionless and predictive digital online shopping will eventually become the norm. Retail stores will standardize AI-enabled smart shelves that control inventory, display dynamic pricing and direct shoppers to related items.

Behind the scenes, AI will help retail organizations further cut waste and lower their carbon footprints by better managing inventory and supply chains.

How long will we have to wait for our new AI-powered shopping experience? At the rate things are moving these days, not long at all.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Looking for business benefits from GenAI? Supermicro, AMD & PioVation have your solution

Featured content

Looking for business benefits from GenAI? Supermicro, AMD & PioVation have your solution

Struggling to deliver business benefits from Generative AI? Supermicro, AMD and PioVation have a new solution that not only works out-of-the-box, but is also highly scalable.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Experimenting with Generative AI can be fun, but CEOs and corporate boards aren’t interested in fun. They want to see real business results—things like an enhanced customer experience, more innovative products, streamlined operations and lower TCO. And they want to see them now.

Getting GenAI to deliver these kinds of business results isn’t easy. A recent report from MIT finds that despite nearly $40 billion worth of enterprise investment in GenAI, 95% of the organizations are getting “zero return.”

That estimate is based on solid numbers. The MIT researchers reviewed over 300 AI projects, interviewed with more than 50 organizations, and surveyed some 150 senior leaders.

The latest forecasts aren’t much cheerier. Research firm Gartner this summer predicted that by the end of this year, nearly a third of all GenAI projects (30%) will be abandoned after the proof-of-concept stage. Gartner says the projects will be cut due to poor data quality, inadequate risk controls, escalating costs and unclear business value.

“After last year’s hype, executives are impatient to see returns on GenAI investments,” says Gartner analyst Rita Sallam. “Yet organizations are struggling to prove and realize value.”

That’s About to Change

Supermicro, AMD and startup PioVation have partnered to jointly develop a GenAI solution that offers a pre-validated, turnkey infrastructure for deploying large language models (LLMs). The benefits include lower deployment overhead, enhanced observability, and ensured control of sovereign data.

Partner PioVation is a developer of AI platforms for enterprises, government agencies, and small and midsize businesses. Its products can be run either on-premises or in PioVation’s cloud in Munich, Germany. The company, founded in 2024 by former AMD executive Mazda Sabony, has formed partnerships with several companies, including AMD and Supermicro.

The GenAI solution being offered by the three companies has been designed to scale all the way from compact on-prem clusters up to large-scale multi-tenant cloud environments. And its architecture integrates Supermicro rack-level systems, AMD Instinct GPUs, and PioVation’s agentic AI platform, PioSphere. The result, the companies say, is out-of-the box agentic AI at any scale.

Full Stack

The Supermicro-AMD-PioVation offering is a full-stack solution. An autonomous microservice chains LLM prompts, invokes domain-specific tools, and integrates seamlessly with your existing systems via REST (an architectural style for distributed hypermedia systems), gRPC (a remote procedure call framework), or event streams running on the pre-validated Supermicro server powered by AMD Instinct GPUs.

Another feature is the solution’s Model Context Protocol (MCP). It lets agents interact with external tools in a way that’s both modular and composable. The MCP also governs how tools are registered, discovered, invoked and composed dynamically at runtime. This includes input/output serialization, maintaining execution context, and enforcing consistency across tool chains. MCP also enables context-aware tool usage, making every agent interoperable, auditable and enterprise-ready from the start.

The solution is available in three topologies, each designed for different operational scales and use cases:

  • MiniStack: For SMBs, pilots, research and the edge.
  • EdgeCluster: For regulated sites, branches and other locations where high availability is required.
  • Cloud Deployment: For cloud service providers (CSPs), enterprises and AI providers.

All three versions include a unified agent dashboard, role-based access control, and policy enforcement.

Business Benefits

The three partners haven’t forgotten about the need for GenAI to deliver real business results that can keep CEOs and corporate boards happy. To that end, the solution offers benefits that include:

  • Turnkey deployment: PioSphere’s Cloud OS has been prevalidated on the Supermicro platform powered by AMD GPUs.
  • Unified operations stack: A tightly integrated environment eliminates fragmented AI tooling.
  • No-code agent development: A PioVation feature known as AgentStudio lets nontechnical users design, deploy and iterate AI agents using a no-code interface.
  • Sovereign data control: Built-in controls support national and regional compliance frameworks, including Europe’s GDPR and the United States’ HIPAA.
  • Multi-tenant scalability: An organization can create separate, secure environments for different business units or clients, yet they’ll all share a common infrastructure footprint.
  • Integrated LLM operations and agent life-cycle management: Users can integrate any LLM published on the Hugging Face or Kaggle communities with one-click connectors. Other built-in features include RAG (retrieval augmented generation) pipelines and full agent life-cycle tools.
  • Intelligent autoscaling: During workload spikes, the solution’s dynamic autoscaling ensures resource utilization, cost efficiency and seamless performance.

Put it all together, and you have a solution that goes far beyond mere experimentation. The three partners—Supermicro, AMD and PioVation—are serious about helping your GenAI projects deliver serious benefits for the business.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Tech Explainer: What’s a short-depth server?

Featured content

Tech Explainer: What’s a short-depth server?

Do your customer have locations that need server compute power, but lack data centers? Short-depth servers to the rescue!

Learn More about this topic
  • Applications:
  • Featured Technologies:

There are times when a standard-sized server just won’t do. Maybe your customer’s branch office or retail store has space constraints. Maybe they have concerns over portability. Or maybe their sustainability goals demand a solution that requires low power and efficient cooling.

For these and other related situations, short-depth servers can fit the bill. These relatively diminutive boxes are designed for use in less-than-ideal physical spaces that nevertheless demand high-performance IT infrastructure.

What kinds of organizations could benefit from short-depth server? Consider your local retail store. It’s likely been laid out using a calculus that prioritizes profit per square inch. This means the store’s best spots are dedicated to attracting buyers and generating revenue.

While that’s smart in terms of retail finance, it may not leave much room for vital infrastructure. That includes the servers that power the store’s point of sale (POS), security, advertising and data-collection systems.

This is a case where short-depth servers can help. These systems provide high levels of compute, storage and networking—without needing tall data center racks, elaborate cooling systems or other supporting infrastructure.

Other good candidates for using short-depth servers include remote branch offices, telco edge installations and industrial environments. In other words, any location that needs enterprise-level servers, but is short on space.

Small but Mighty

What’s more, today’s short-depth servers can handle some serious workloads.

Consider, for instance, the Supermicro WIO A+ Server (AS -1115SV-WTNRT), powered by AMD EPYC 8004 series processors. This short-depth server is engineered to tackle a variety of workloads, including virtualization, firewall applications, database, storage, edge and cloud computing.

The WIO A+ ships as a 1U form factor with a depth of just 23.5 inches. Compared with one of Supermicro’s big 8U multi-GPU servers, which has a depth of more than 33 inches, the short-depth server is short indeed.

Yet despite its diminutive size, this Supermicro server is packed with a ton of power—and room to grow. A single AMD EPYC processor sits at the center of the action, aided by either one double-width or two single-width GPUs.

This server also has room for up to 768GB of ECC DDR5 memory. And it can accommodate up to 10 hot-swap drives for NVMe, SAS or SATA storage.

As if that weren’t enough, Supermicro also includes room in this server cabinet for two PCIe 5.0 x16 full-height, full-length (FHFL) expansion cards. There’s also space for a single PCIe 5.0 x16 low-profile (LP) card.

More Power for Smaller Space

Fitting enough tech into a short-depth server can be a challenge. To do this, Supermicro’s designers had a few tricks up their sleeves.

For one, they used a custom motherboard instead of the more common ATX or EEB types. This creates more space in the smaller chassis. It also lets the designers employ a high-density component layout. The processors, GPUs, drives and other elements are placed closer to each other than they could be in a standard server.

Supermicro’s designers also deployed low-profile heat sinks. These use pipes that direct the heat toward fans. To save space, the fans are smaller than usual, but make up the difference by running faster. Sure, faster fans can create more noise. But it’s a worthy trade-off to avoid system failure due to overheating.

Are there downsides to the smaller form factor? There can be. For one, constrained airflow could force a system to throttle both processor and GPU performance in an effort to prevent heat-related issues. This could be an issue when running highly resource-intensive VM workloads.

For another, the smaller power supply units (PSUs) used in many short-depth servers may necessitate a less-powerful configuration than a user might prefer. For example, Supermicro’s short-depth server includes two 860-watt power supplies. That’s far less available power than the company’s multi-GPU powerhouse, which comes with six 5,250-watt PSUs. Of course, from another perspective, the need for less power can be seen as a benefit, especially at remote edge locations.

Short-depth servers represent a useful trade-off. While they give up some power and expandability, their reduced sizes can help IT pros make the most of tight spaces.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Pages