Sponsored by:

Visit AMD Visit Supermicro

Performance Intensive Computing

Capture the full potential of IT

Tech Explainer: Why embedded systems for retail?

Featured content

Tech Explainer: Why embedded systems for retail?

You’ll find embedded systems in thousands of retail locations—but only if you know where to look. Find out how these specialized servers work, why they make sense, and how your retail customers can get started using them.

Learn More about this topic
  • Applications:
  • Featured Technologies:

While dedicated high-performance servers and multiuse cloud platforms command the biggest headlines in tech news, that doesn’t mean they’re the perfect fit for every use case.

Retail organizations have unique requirements. Sometimes those requirements are best served by the diminutive, unsung heroes of the server world: embedded retail servers.

Embedded systems are usually smaller and less powerful than their larger, purpose-built cousins. So where giant AI servers may offer the brute force power of a freight train, smaller embedded systems are more like a ski lift. They do only one thing, but they do it very well.

You can find embedded systems in thousands of retail sites, but you’ll have to do some hunting—their location is not always obvious. Some embedded retail servers sit under counters or in small, out-of-the-way closets. Others are attached to the backs of large color displays that offer patrons dynamic menus, ads and special deals.

High-Tech Sales

One of the most common retail embedded systems is the humble point-of-sale (POS) terminal. A quick survey of your favorite retail stores is likely to reveal a variety of versions, ranging from smart cash registers to fully autonomous self-checkout kiosks.

But POS devices are designed to do far more than just add prices and calculate tax. In a modern retail setting, these servers may also read barcodes, weigh items, process mobile payments, update inventory, schedule deliveries, and detect fraud.

These processes can become even more demanding when the embedded system must complete them without the aid of cloud services.

Why? Because without the processing power and storage of remote cloud and core servers, the embedded system has to rely on its own internal components to complete what can often be a series of very demanding tasks.

Other Use Cases

Deploying embedded retail systems becomes even more complex when a retail location doubles as a warehouse. Such is the case with supermarkets and big-box retailers like Walmart. They must be able to quickly restock their shelves whenever supplies are depleted by shoppers.

In these locations, you can often find embedded retail servers keeping track of real-time stock levels. This can be accomplished using a number of methods, including radio frequency identification (RFID) tags, shelf-based weight sensors, and AI-enabled cameras.

Another task best handled by small, embedded systems is building and energy management. Retail operations often use embedded servers connected to distributed sensors to control HVAC, lighting and security. Here, again, it’s vitally important that these systems be able to operate without an internet connection when necessary.

In this case, an embedded server’s ability to operate on its own can actually prevent physical disasters. Even deprived of remote cloud services, it may need to keep control over a store’s climate to prevent damaging stock. Likewise, store managers often rely on an embedded server’s ability to maintain 100% security system uptime to avoid theft, damage or fire.

Power to Get the Job Done

Designers of embedded retail servers have a tricky job. They need to create systems that meet a long list of disparate requirements. That’s because the most effective embedded retail servers are:

  •  Compact enough to fit in small retail outlets
  • Cost-effective enough that enterprises can outfit each location with multiple servers
  •  Powerful enough to handle multiple complex tasks and run AI applications locally
  • Outfitted with enough storage to collect terabytes of data
  • Reliable enough to run security services that store managers can rely on 24x7
  • Able to reliably perform with or without an internet connection

To address these concerns, systems designers like Supermicro are tasked with creating the perfect balance of power, pricing, and reliability.

One such well-balanced embedded server is the upcoming Supermicro IoT A+ Server (AS-E300-14GR). It’s a mini-1U server powered by AMD EPYC 4004/4005 series processors with 16 cores and a 64MB cache.

Despite its small size, Supermicro’s embedded system still manages to offer some real expansion. For example, users can have Supermicro populate the server with up to 960GB of SSD storage and 192GB of DDR5 RAM. They can also opt for additional storage via the system’s dual M.2 PCIe 5.0 x4 NVMe slots.

In addition, there’s a single PCIe 5.0 x16 LP slot for an expansion card. Common options to fill that slot include PCIe-based networking cards and dedicated AI accelerators like AMD’s Instinct GPUs.

Coming Soon

What kind of features can we expect from future generations of embedded retail servers? The answer will have much to do with consumer shopping habits, economic and market shifts, and new tech that becomes available in the near future.

While we can’t make infallible predictions about those forces, we can make some assumptions.

One is that AI will become deeply integrated in embedded systems. In fact, we could soon see more systems with AI fully on-device—no cloud connection necessary.

Connectivity-wise, future embedded systems could feature not only Wi-Fi 7 integration, but also 5G cellular connections.

Embedded systems’ footprints should also shrink, even as they become more powerful. Ultra-low-power chips should enable them to operate silently with passive cooling systems and improved thermal management, which will allow designers to shrink the server’s overall size.

Bottom line: Expect embedded servers for retail to become smaller, faster and better. Isn’t that always the way when it comes to new technology?

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Meet Supermicro’s latest MicroBlade, powered by AMD EPYC 4005 processors

Featured content

Meet Supermicro’s latest MicroBlade, powered by AMD EPYC 4005 processors

This Supermicro/AMD platform offers a flexible, density-optimized blade architecture for a 6U system. Pack it to the max, and you’ll get 320 server nodes in a standard 48U rack.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Supermicro has introduced a MicroBlade platform powered by the latest AMD EPYC 4005 series processors. Its intended workloads include cloud, virtualization, data services, edge computing and Software as a Service (SaaS).

These MicroBlades are offered in a new 6U system that supports up to 20 blades in a single enclosure. Each blade holds two CPU nodes, so a fully loaded system with 20 blades has 40 nodes. Fill a standard 48U rack with eight of these 6U systems, and you get a total of 160 blades with 320 server nodes.

The Supermicro system also lets customers mix newer processors with older ones. So customers can expand and upgrade only when their compute requirements change, protecting their earlier investments and providing seamless scalability.

Blade Power

Supermicro MicroBlades offer a powerful and flexible extreme-density 3U and 6U all-in-one blade architecture. Compared with standard 1U rackmount servers, they provide up to 86% more power efficiency and up to 56% improved density, Supermicro says.

The new Supermicro MicroBlade (model number MBA-315R-1DE12) is a dual-node device that measures roughly 23 x 5 x 1 inches and weighs just over 3 pounds. In addition to the two AMD processors, the MicroBlade packs in dual 25GbE LAN ports with 100G uplinks for high-speed networking. There are also two slots for up to 128GB of DDR5 memory.

The blade supports the Intelligent Platform Management Interface (IPMI) v.2.0 via Chassis Management Module. CMM offers remote control of individual server blades, power supplies, cooling fans and networking switches. This lets sys admins control maximum power consumption, manage power allocation, reboot and reset the server, and obtain BIOS configuration data—all remotely with a processor that operates independently of the managed systems.

AMD EPYC 4005 series processors are designed for entry-level systems used by small businesses and hosted IT services providers. While affordable, they offer high performance, advanced technologies and energy efficiency.

Do you have small-business or IT service provider customers looking for a flexible yet powerful server? Tell them about the new Supermicro MicroBlade platform powered by AMD EPYC 4005 series processors.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Tech Explainer: What’s a Neocloud?

Featured content

Tech Explainer: What’s a Neocloud?

This cloud variant has arisen to meet the needs of AI developers. Find out how it differs from hyperscalers—and why your customers might want to jump on board.

Learn More about this topic
  • Applications:
  • Featured Technologies:

A new kind of technology demands a new kind of cloud.

Sure, it’s easy to take cloud computing for granted. After all, it’s been years since “the cloud” became part of our lives and everyday vernacular.

Over the years, clouds ranging from the simple (think Dropbox) to the fabulously complex (think multicloud ecosystems) have been powerful enough to handle whatever we’ve thrown their way.

But now our widespread adoption of AI demands a new kind of cloud.

To the rescue: Behold the neocloud!

Neoclouds offer AI workload-specific functionality as a service. And to help save enterprises and SMBs considerable expenses of time and money, neoclouds offer platforms designed to empower the rapid development and launch of the latest AI creations.

A neocloud isn’t your typical “run anything” platform. Instead, it’s optimized to run a narrow selection of highly specialized AI-centric tasks. These include AI/ML inference and training, data analytics and media rendering.

Neoclouds vs. Traditional Clouds

To better understand how neoclouds fit into the grand scheme of modern cloud architecture, it helps to compare and contrast them with their forebearer, the hyperscaler.

Hyperscalers that include Amazon Web Services (AWS), Microsoft Azure and Google Cloud also offer cloud-based services. They simply offer a much larger and less AI-specific selection.

The seemingly endless array of services these hyperscalers offer makes them ideal for developers who prize flexibility and versatility. Hyperscalers let developers combine multiple managed services to simultaneously harness the power of distributed databases, machine-learning pipelines and other components of a highly customized platform.

By contrast, neoclouds are tuned for specific workloads. They offer a narrower focus and so-called “opinionated architecture” designed to make autonomous architectural decisions. That level of specificity and autonomy changes the nature of the development process from DIY to plug & play.

 

                  

 

More-Specific Hardware, Too

To fully compare neocloud apples with hyperscaler oranges, you also need to look under the hood. The tech behind the latest cloud type makes a huge difference.

For both hyperscalers and neoclouds, we’re talking about some of the most advanced tech ever. But here again, it’s the neocloud’s laser-like focus on AI that makes it an invaluable development tool.

It’s for that reason that popping the top off an AI server like the Supermicro’s 8U server (model AS -8126GS-TNMR) will treat you to a view of truly cutting-edge CPUs, GPUs and networking gear. That gear includes a couple of server-focused AMD EPYC 9005 series processors with as many as 384 cores and up to 6TB of DDR5 memory.

For brute-force AI processing, the Supermicro A+ server also offers room for eight onboard AMD Instinct MI350X GPUs banded together via AMD Infinity Fabric Link.

Supermicro’s behemoth is also equipped with AMD ROCm. Pronounced “rock-em,” it’s a software stack designed to translate the code written by programmers into sets of instructions that AMD GPUs can understand and execute perfectly.

The Neocloud Sales Pitch, Condensed

The what and how of neoclouds are important. But if your customers are considering investing in neocloud, they’ll surely want to know about the why, as well.

So why would you want to engage a neocloud for AI development? There are four main reasons:

1. Neoclouds cut admin work, letting you concentrate instead on production.

A new eBook from Supermicro and AMD, The Smartest Path to Scalable AI, cites neoclouds for their “frictionless dev-to-prod motion.”

That’s tech business-speak for a system that handles the nitty-gritty details, getting out of your way so you can get to work. That includes one-command access to optimized hardware and preconfigured environments.

Bottom line: Less admin, more development, and faster time-to-market.

2. A neocloud delivers instant gratification, not endless development integration.

“Day 0 readiness” is the catchphrase that sums up this one. And not just for any single aspect of the neocloud platform, but for the whole stack. That includes hardware, software, and the managed offerings wrapped around them, collectively referred to as services.

Bottom line: Large models and agents start running efficiently from the get-go.

3. A neocloud is always up-to-date with the latest, greatest silicon.

The last thing you want to contend with is outdated infrastructure. That may fly when it comes to making last-decade file storage app. But creating tomorrow’s brilliant new AI requires cutting-edge tech. The problem is, that tech gets expensive. The solution? Rent, don’t buy.

Bottom line: Access to all the cool toys, with no down payment.

4. It’s already got wheels; you don’t have to reinvent them.

Neoclouds come well stocked with what are known as specialized microservices. These are pre-built, workload-specific building blocks that developers can stand on to bypass the mundanities of production and get to the good stuff.

Examples of wheels you won’t have to reinvent include distributed training orchestration, streaming ingestion services, and GPU render farms.

Bottom line: Neoclouds do the boring due diligence, and let developers get all the glory.

The Future’s Future

Neoclouds are already the future. They’re coming online now, and revealing themselves to be the greatest thing for developers since sliced bread.

But tech moves fast these days. There’s always someone thinking about the next step.

When it comes to the next step for neoclouds, that’s likely to involve deeper specialization, more compelling economics, and consolidation.

That makes sense in terms of the big picture. As both enterprises and SMBs adopt neoclouds, they’ll create more demand. That demand, in turn, should help fund expansion.

Eventually, we may see a new level of specificity. For example, one neocloud could offer low-latency SaaS production inferencing, while another may focus on analytics that cater to medical research.

What happens after that is hard to predict. But one easy-to-believe theory foretells a time in which neoclouds plug into hyperscalers. With that kind of power, imagine what tomorrow’s developers will be able to do!

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Need to cool AI hardware with safety? Check out the new solution from AMD, Supermicro & Metrum AI

Featured content

Need to cool AI hardware with safety? Check out the new solution from AMD, Supermicro & Metrum AI

The solution employs AI agents to monitor liquid-cooling systems, identifying and remediating problems quickly.

 

Learn More about this topic
  • Applications:
  • Featured Technologies:

Liquid cooling is great for controlling the temperature of hard-working AI servers, but the technology also has its risks. Even minor disruptions or fluctuations in a cooling system can quickly lead to massive hardware failures.

A solution to this challenge has been developed by AMD and Supermicro working with Metrum AI Inc., an Austin, Texas-based provider of industry-specific AI agents and AI evaluation products.

Their solution integrates Supermicro infrastructure, AMD computational power and ROCm software, and Metrum AI’s orchestration to deliver fast decisions that ensure safety at scale.

This solution enables multiple AI agents to collaboratively monitor signals, diagnose issues, predict failures, and coordinate corrective actions. The agents are embedded directly into a server’s cooling control plate.

Essentially, this creates a data center that is adaptive, resilient, and self-optimizing. The solution should also support the massive compute intensity of next-generation AI workloads while proactively managing their thermal and physical risks.

And unlike traditional monitoring tools, this solution can actually predict and then prevent catastrophic hydraulic failures—before they occur. And do so faster than would be possible with traditional human intervention.

Power Features

To design these multi-agent systems, the team used AMD ROCm. This open-source software delivers important benefits that include flexibility, optimized libraries and seamless integration with AMD Instinct GPUs.

Another feature that made the solution possible is the massive memory reservoir of the AMD Instinct GPUs. For example, the AMD Instinct MI355X GPU has a dedicated memory of 288 GB. This lets large-scale reasoning models operate fully in-memory.

The structural foundation of this platform is the Supermicro 8U server (model AS -8126GS-TNMR) powered by dual AMD EPYC 9005 Series CPUs and supporting up to eight AMD Instinct MI325X or MI350X GPUs.

Unlike standard servers, these systems are engineered with direct-to-chip cooling headers that expose flow, temperature and pressure data directly through Redfish interfaces. (Redfish is a standard designed to deliver simple and secure management for converged systems, hybrid IT and software-defined data centers.) This empowers the agents to monitor and adjust cooling performance in real time.

The combination of specific technologies creates what’s known as a Unified Computational Fabric. There, the AMD EPYC processors feed continuous Redfish data directly into the Supermicro Instinct GPUs via PCIe 5, eliminating I/O bottlenecks.

This synergy powers the platform to sustain real-time adaptive control loops across dozens of racks, and quickly. It’s a capability that conventional air-cooled and CPU-based infrastructures can’t deliver.

Smart Racks

The autonomous cooling system was built on a distributed multi-agent architecture designed specifically for liquid-cooled AI environments. Unlike conventional systems, where monitoring is either centralized or based on human intervention, the solution places intelligence directly at the rack level.

In this setup, lightweight agents continuously monitor telemetry, interpret changes in flow and pressure, and coordinate rapid remediation actions across the data center. This creates a resilient, high-resolution control fabric that can respond to thermal events in milliseconds.

At the base of the stack, AMD ROCm supplies the core libraries, tools, compilers and runtimes for GPU-accelerated compute on AMD Instinct GPUs. And Kubernetes orchestration and the AMD GPU Operator enable containerized deployment, GPU scheduling, and lifecycle management at a multirack scale. (Kubernetes is an open-source system for automating the deployment, scaling and management of containerized applications.)

Above this layer, the AMD Enterprise AI Suite delivers higher-level services. The suite is a full-stack of enterprise-ready AI. The services it delivers include solution blueprints, AI workbench, and a resource manager for unified model deployment, optimization and infrastructure governance.

Metrum AI extends these platform components into a specialized multiagent architecture. It supports real-time telemetry ingestion, large-model reasoning and autonomous cooling control.

 

 

 

Test Results: Fast Yet Stable

All that sounds good in theory, but does it really work?

To find out, the solution was tested by Metrum AI along two dimensions: telemetry ingestion thruput and large-model inference stability.

When monitoring a full deployment of 200 racks (1,000 servers), the system successfully processed more than 13,000 Redfish telemetry endpoints per minute. Simultaneously, it maintained over 8,000 tokens/second of multiagent large-model reasoning.

This demonstrated that as the infrastructure added complexity, the centralized coordination architecture did not become a bottleneck. Also, the test shows that every agent received real-time, high-resolution sensor context, regardless of facility size.

Across all benchmarks, the integrated solution demonstrated stable, real-time, end-to-end autonomous operation under data-center scale load.

So do you have customers who are eager to try liquid cooling, but concerned about the risks? If so, tell them about this new AI-powered solution from AMD, Supermicro and Metrum AI.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Meet Supermicro’s new AMD-powered edge systems

Featured content

Meet Supermicro’s new AMD-powered edge systems

The five new systems range from compact systems to rackmount servers. They’re designed for use outside of the traditional data center.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Supermicro has five new edge systems, all powered by AMD EPYC processors.

The new hardware are components of Supermicro’s end-to-end portfolio of edge systems. They represent three system categories: compact edge systems, compact edge servers and rackmount edge servers. All are designed for use at the edge—that is, outside the traditional data center.

Supermicro describes the new servers as “purpose-built solutions for a scalable, performance-efficient edge infrastructure.” To that end, the company is offering different configurations for different workloads.

For AI workloads, Supermicro is offering GPU-optimized platforms. For sensor and network integration, there are flexible I/O offerings. For harsh environmental and industrial settings, fanless designs. And for mobile deployments and small spaces, there are compact form-factor designs.

The new edge servers are designed for users in industries that include healthcare, retail, telecom and manufacturing. All share the need for reliable compact systems.

Here’s a look at the five new servers, arranged by system category.

Category 1: Compact edge systems

Supermicro’s compact edge systems are designed to run specialized workloads and connect on-prem equipment to enterprise networks. They can include both fanless gateways built for harsh industrial settings and high-performance edge servers running advanced AI models.

Organizations that deploy edge systems can reduce latency, lower bandwidth costs, improve security, and ensure continuous operations for mission-critical applications.

Supermicro’s new entry in this category is a mini 1U embedded system, model number AS -E300-14GR. It’s powered by the user’s choice of a single AMD EPYC 4004 or 4005 processor with up to 16 cores and 32 threads.

This air-cooled system measures 15 x 10.9 x 5.6 inches, and it weighs just 7.5 pounds.

Intended applications include healthcare, surveillance, AI inference, digital signage and point-of-sale. For AI inference, the system also supports up to 4 GPU accelerator cards.

Category 2: Compact edge servers

These edge servers combine specialized features with expansion options that include support for GPU accelerators. They’re designed for AI inference in industries such as retail, manufacturing and smart spaces.

For this category, Supermicro offers 2 new AMD-powered servers.

The first is a 2U compact AI system, model number AS -2116S-TNRT. It’s powered by a single fifth-generation AMD EPYC 9005 series processor with up to 192 cores, 384 threads and 384MB of cache. This server supports up to 1 double-width or 5 single-width GPUs. And it’s air-cooled by 4 fans.

The other new system is a compact 1U short-depth server, model number AS -1115S-FWTRT. This server measures 17.2 x 16.9 x 1.7 inches, and it weighs 15 pounds.

This short-depth server is powered by a single AMD EPYC 8004 processor. And it’s air-cooled by up to 6 fans.

The system has been designed for applications that include virtualization, firewall, cloud services, CDN (content delivery network), 5G networks, and VRAN/ORAN (virtual and open radio access network) for telecom.

Category 3: Rackmount edge servers

These are mainly short-depth servers designed for embedded, edge and telecom workloads that include RAN and edge AI. They offer a modern combo: high-density compute plus GPU compatibility.

For this category, Supermicro offers two AMD-powered systems. The first is a 2U Hyper-E server, model AS -2115HE-FTNR. This server is powered by a single AMD EPYC 9004/9005 series processor with up to 160 cores and 320 threads.

The box can accommodate up to 3 double-width GPUs. And it’s packed with up to 6TB of DDR5 memory. For extra flexibility, the server is available in both front and rear I/O models.

For this category, Supermicro also recently introduced a ultra-short-depth, low-power edge and embedded platform, model number AS -116R-FNR. It measures 17.2 x 9.8 x 1.7 inches.

This system is powered by a single AMD EPYC 4005 processor. And it can handle up to 192GB of DDR5 memory.

Have clients looking for compact yet powerful systems and servers they can run outside the traditional data center? Tell them about the new AMD-powered Supermicro edge.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Tech Explainer: What’s an AI Factory?

Featured content

Tech Explainer: What’s an AI Factory?

Discover how AI factories work—and how your clients might benefit from building an AI factory of their own.

Learn More about this topic
  • Applications:
  • Featured Technologies:

How can you tell that the AI Era is here? One way is by noticing that large enterprises are increasingly focused on mass producing AI models.

It’s no longer enough to have a decent set of working AI models to power Spotify’s suggestion engine or Accenture’s Big Data analytics.

To keep up with—and surpass—the Joneses, Spotify and Accenture will need dedicated systems that work every day to create, evaluate and iterate their AI models.

These systems are called AI factories. Somewhat like a factory that creates physical widgets, an AI factory churns out new and updated AI models. This continual AI production process helps enterprises react quickly to market demands and competition.

Make no mistake: The development of AI factories represents a turning point in the evolution of AI-powered business.

No. 2 with a Bullet

This theory is supported by some of IT’s top thinkers. They include Tom Davenport, a professor, speaker and author; and Randy Bean, a corporate advisor.

Davenport and Bean co-wrote an article that appeared earlier this month In the Sloan Management Review: Five trends in AI and data science for 2026. In their article, the authors place AI factories in the Number 2 spot. AI factories, they say, will be adopted by users and “all-in” AI adopters that include consumer products makers, banks and software companies.

As Davenport and Bean explain, an AI factory combines technology platforms, methods, data and previously developed algorithms to make building AI systems easy and fast. The authors’ all-important message: Watch this space.

How AI Factories Work

To fully understand the concept of an AI factory, it can help to think of the traditional smoke-belching, brick-and-mortar factories it’s named for.

Of course, there are some differences. A physical factory takes in raw materials, uses machines to process them, and produces physical products.

By contrast, an AI factory takes in data (such as text, audio, images and logs), runs that data through massive compute engines, and outputs AI models for recommendations, predictions, automation and generative content.

Another difference: Unlike the static products that emerge from traditional factories, the products of AI factories are virtual. They learn and grow as new data, infrastructure and techniques become available. In this way, AI factories help their organizations keep up with rapid changes and market shifts.

For instance, a new AI model produced by an enterprise’s AI factory can be continuously retrained as new data becomes available. While each new iteration deployed in the field busily suggests which Netflix movie to watch next, a newer version is constantly being developed in the background. When the new suggestion engine is ready, Netflix can seamlessly slide it into place.

Why Your Clients Probably Need an AI Factory

It’s good to understand the abstract benefits of an AI factory. But your clients will also want to know how building one can translate into business results.

Here’s the bottom line. An AI factory can:

  • Dramatically reduce the cost of business intelligence. Once an AI factory is built and a given AI model is trained, that model can run continuously, serving millions of decisions, predictions, etc., for a fraction of its initial cost. In other words, the cost per additional decision rapidly collapses toward zero.
  • Help organizations maintain a decisive competitive advantage. This happens on two levels. First, maintaining a constant production stream of AI models and iterations helps your clients meet market demands as quickly as possible. And second, having that ability to react faster to customer needs and economic conditions can help create and sustain an advantage over competitors.
  • Turn data into capital. Many organizations are ill-equipped to analyze and monetize all the data they collect. All that piled-up data can seem like an albatross around their neck. But by building an AI factory, the organization can harness that otherwise squandered data and put it to work.

Further, companies that don’t build an AI factory could find themselves at a competitive disadvantage. Davenport and Bean, in their Sloan Management Review article, say companies that lack an AI factory will find building AI at scale both expensive and time-consuming.

Stumbling Blocks? A Couple

Building an AI factory isn’t always easy. Enterprises can run into serious roadblocks.

For one, siloed, inconsistent or low-trust data can make for a messy AI production process. As programmers say, “garbage in, garbage out.” In other words, if the data is messy, the analysis will be, too.

Another thing that can wreak havoc on the virtual factory floor are talent bottlenecks. There are only so many data scientists to go around, and they’re in high demand. Finding the right employees is a key component here—even in an age of super-smart robots.

Another trap your clients need to watch out for are bureaucratic hold-ups. Legal, compliance and trust issues can cause AI projects to grind to a halt.

The AI Factory Future

Like everything else in the fast-moving AI world, AI factories are changing. In the near future, AI factories will likely focus on the immediacy of real-time, always-on learning.

As AI factories shift to nearly continuous adaptation, enterprises will use their AI model updates to keep pace with rapidly changing market conditions and customer demands.

Another likely future is inferencing at the edge. For “edge,” think vehicles, devices and brick-and-mortar factories. Organizations that move inferencing closer to where data is created can lower system latency (that is, increase speed) and reduce cloud costs.

Another factor that could make a big impact on AI factories is new software and hardware integrations. A recent Supermicro webinar on AI factories and related technology showed how enterprises can benefit from integrating software platforms such as Supermicro’s SuperCloud Composer (SCC) and Power Asset Orchestrator (PAO).

Supermicro says this potent combination allows operators to gain total visibility into AI Factories. It can also optimize everything from GPU telemetry to real-time grid pricing.

Overall, it’s safe to assume that when these and other updates are deployed, AI factories will quickly become part of the common AI infrastructure. In so doing, they’ll touch nearly every aspect of our daily lives.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

i3D Supercharges Game Hosting with AMD EPYC Processors

Featured content

i3D Supercharges Game Hosting with AMD EPYC Processors

Discover how game host i3D boosted per-core performance, improved the user experience, and improved its TCO…all by moving to Supermicro MicroBlade servers with AMD EPYC 4000 Series processors.

Learn More about this topic
  • Applications:
  • Featured Technologies:

One thing hosts of fast-moving multiplayer games don’t want is jitter.

Jitter is the game industry’s term for an inconsistent user experience. It occurs when what’s known as the tick rate base workload—the frequency with which gamers are updated—differs from one game-player to another.

Keeping this tick rate consistent across gamers isn’t easy. Some games get updated hundreds of times a second.

That explains why global game hosting provider i3D.net recently refreshed its infrastructure stack. The company chose Supermicro MicroBlade servers powered by AMD EPYC 4004 Series processors.

Single-Core Rules

To understand i3D’s choice, it helps to understand how the demands of game hosting differs from those of conventional cloud hosting.

Cloud hyperscalers generally try to pack as many compute cores as possible into the smallest possible space. That’s because they want to support many virtual machines on a single node. To get this result, they buy large servers with lots of cores and plenty of memory.

By contrast, for game hosting providers, it’s single-core performance that rules. These companies want to provide the best possible performance for their users. For this purpose, core count per CPU is relatively unimportant.

One thing that really matters for game hosting is single-core performance. And to control costs, gaming providers typically scale with lots of smaller machines rather than a single big one.

All that’s important for i3D. The company, founded in the Netherlands in 2002, initially rented consumer game servers. In 2018, i3D was acquired by Paris-based Ubisoft, and today it offers not only game online services, but also cloud and compute resources, connectivity services, and colocation via its private data center.

Big Games, Big Systems

i3D planned its rollout in large part to support a new game, “Dune: Awakening.” It’s a massively multiplayer online game.

To provide the needed scale, i3D acquired Supermicro MicroCloud servers powered by AMD EPYC 4464P processors. This CPU, part of AMD’s 4th generation EPYC 4004 series, packs 12 cores, 24 threads and 64MB of cache. Yet its power consumption is just 65 watts, a level that fits most data centers.

Now that the rollout is complete, i3D has found that single-core performance with the new setup on a bare-metal is 52% higher than the previous solution.

As Paul Louvet, i3D’s senior product manager of bare metal, puts it: “AMD has the best performance out there for a very attractive cost.”

Double the Nodes

These AMD processors power i3D’s choice of Supermicro 3U MicroCloud servers with eight nodes (Supermicro model AS -3015MR-H8TNR). Each node has a single AMD CPU. This means i3D can fit 96 nodes in one rack, more than double what they could do before.

The Supermicro chassis also includes dual power supplies, bolstering reliability.

Though the upgrade involved a transition from older servers based on a competitor’s processors, i3D says the shift to AMD was seamless. I3D now has about 1,800 nodes powered by AMD EPYC processors, and it will add even more soon.

Looking ahead, i3D plans to upgrade the Supermicro servers to AMD EPYC 4545P processors. This CPU is a member of AMD’s 5th generation EPYC 4005 series, which offers ‘Zen 5’ data-center processors designed for small businesses and hosted services.

Importantly, these processors offer 16 cores to the prior generation’s 12. That will allow i3D to employ four additional cores for the same power usage.

“That’s incredible,” says Louvet of i3D. “This CPU will allow us to have much better TCO per core.”

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

2025: Look Back at the Year’s Top Advances

Featured content

2025: Look Back at the Year’s Top Advances

Catch up on 2025’s highlights: ROCm 7.0, liquid-cooled AI servers, server processors for SMBs, and a MicroBlade server that’s highly efficient.

Learn More about this topic
  • Applications:
  • Featured Technologies:

2025 was a year to remember. But in case you’ve forgotten, here are some of the year’s top advances.

ROCm for the AI Era

This past fall, AMD introduced version 7.0 of its ROCm software stack. This latest edition features capabilities designed especially for AI.

ROCm, part of AMD’s portfolio since 2016, translates code written by human programmers into instruction sets that AMD GPUs and CPUs can understand and execute.

Now AMD has purpose-built ROCm 7.0 for GenAI, large-scale AI training, and AI inferencing. Essentially, ROCm now offers the tools and runtime to make the most complex GPU workloads run efficiently.

The full ROCm 7.0 stack contains multiple components. These include drivers, a Heterogeneous Interface for Portability (HIP), math and AI libraries, compilers and system-management tools.

Liquid-Cooled AI Servers

Supermicro introduced two rackmount AI servers in June, both of them powered by AMD Instinct MI350 Series GPUs and dual AMD EPYC 9005 CPUs.

One of the two new servers, Supermicro model number AS -4126GS-NMR-LCC, is a 4U liquid-cooled system. This server can handle up to eight GPUs, the user’s choice of AMD’s Instinct MI325X or MI355X.

The other server, Supermicro model number AS -8126GS-TNMR, is a larger 8U server that’s also air-cooled. It also offers a choice of AMD GPUs, either the AMD Instinct MI325X or AMD Instinct MI350X.

Both servers feature PCIe 5.0 connectivity; memory capacities of up to 2.3TB; support for AMD’s ROCm open-source software; and support for AMD Infinity Fabric Link connections for GPUs.

In June, Supermicro CEO Charles Liang said the new servers “strengthen and expand our industry-leading AI solutions—and give customers greater choice and better performance as they design and build the next generation of data centers.”

EPYCs for SMBs

In May, AMD introduced a CPU series designed specifically for small and medium businesses.

The processors, known as the AMD EPYC 4005 Series, bring a full suite of enterprise-level features and performance. But they’re designed for on-prem SMBs and cloud service providers who need cost-effective solutions in a 3U form factor.

“We’re delivering the right balance of performance, simplicity, and affordability,” says Derek Dicker, AMD’s corporate VP of enterprise and HPC. 

That balance includes the same AMD ‘Zen 5’ core architecture behind the AMD EPYC 9005 Series processors used in data centers run by large enterprises.

The AMD EPYC 4005 Series CPUs for SMBs come in a single-socket package. Depending on model, they offer anywhere from 6 to 16 cores and boosted performance of up to 5.7 GHz.

One model of the AMD EPYC 4005 line also includes integrated AMD 3D V-Cache technology for a larger 128MB L3 cache and lower latency.

MicroBlades for CSPs

The AMD EPYC 4005 Series processors made a star appearance in November, when Supermicro introduced a 6U, 20-node MicroBlade server (model number MBA-315R-1G) powered by the new CPUs.

These servers are intended for small and midsize cloud service providers.

Each blade is powered by a single AMD EPYC 4005 CPU. When 20 blades are combined in the system’s 6U form factor, the system offers 3.3x higher density than a traditional 1U server. It also reduces cabling by up to 95%, saves up to 70% space, and lowers energy costs by up to 30%.

This MicroBlade system with an AMD EPYC 4005 processor is also available as a motherboard (model number BH4SRG) for use in Supermicro A+ servers.

~~~~~~~~~

Happy holidays from all of us at Performance Intensive Computing, and best wishes for the new year! We look forward to serving you in 2026.

~~~~~~~~~~

Read related 2025 blog posts:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Tech Explainer: What are CPU Cores, Threads, Cache & Nodes?

Featured content

Tech Explainer: What are CPU Cores, Threads, Cache & Nodes?

Today’s CPUs are complex. Find out what the key components actually do—and why, in an age of AI, they still matter.

Learn More about this topic
  • Applications:
  • Featured Technologies:

In the age of artificial intelligence, CPUs still matter. A central processor’s parts—cores, threads, cache and nodes—are as important as any AI accelerator.

But what exactly do those CPU parts do? And why, in an age of AI, do they still matter?

These questions are easy to overlook given AI’s focus on the GPU. To be sure, graphical processors are important for today’s AI workloads. But the humble CPU also remains a vital component.

If the GPU is AI’s turbocharger, then the CPU is the engine that makes the whole car go. As Dan McNamara, AMD’s GM of compute and enterprise AI business, said at the recent AMD Financial Analysts Day, “AI requires leadership CPUs.”

So here’s a look at the most important components of today’s data-center x86 CPUs. And an explanation of why they matter.

Cores: Heavy Lifting

The central processing unit is the brain of any PC or server. It reads instructions, does the complex math, and coordinates the system’s every task.

Zooming into the architecture of a CPU, it’s the individual cores that put the “PU” in CPU. Each fully independent processing unit can run its own task, virtual machine (VM) or container.

Modern enterprise-class CPUs such as AMD’s EPYC 9005 Series offer anywhere from 8 to 192 cores each. They operate at up to 5GHz.

These cores are built using AMD’s ‘Zen’ architecture. It’s a fundamental core design that offers enhancements vital to data centers, including improved instructions-per-clock (IPC), branch prediction, caches and efficiency.

Performance like that is a must-have when it comes to a data center’s most demanding tasks. That’s especially true for compute-intensive database operations and API-heavy microservices such as authentication, payment gateways and search.

Having more cores in each CPU also enables IT managers to run more workloads per server. That, in turn, helps organizations lower their hardware and operating costs, simplify IT operations, and more easily scale operations.

Threads: Helping Cores Do More

A modern CPU core needs to multitask, and that’s where having multiple threads is essential. A single CPU core with two threads can juggle two tasks by switching between them very quickly. In a CPU with a high core count, a productivity-multiplier like that becomes exponentially more effective.

This capability delivers two important benefits. One, it helps ensure that each CPU core stays productive, even if one task stalls. And two, it boosts the CPU’s overall output.

For example, the AMD EPYC 9965 processor boasts 192 cores with a total of 384 threads. That kind of multitasking horsepower helps smooth request handling for web services and microservices. It also improves VM responsiveness and helps AI workloads run more efficiently under heavy loads.

Cache: Speedy but Short-Term Memory

The unsung heroes of CPU design? That would be cache.

The main job of a CPU cache is to help the cores juggle data with low latency. Remember, less latency is always better.

As a result, CPU cache enables databases to run faster, improve VM density and reduce latency.

Your average CPU cache is arranged in three layers:

  • L1 cache is very small and very fast. Each core has its own L1 cache, which holds around 32 KB of instructions and data. The L1 cache sends that data to a register— a tiny, ultra-fast storage location the core uses to acquire the data used for calculations.
  • L2 cache is also exclusive to each core. At around 1MB, this cache is bigger than L1, but it’s also a little slower. L2 cache holds any data that doesn’t fit in the L1 cache. Working together, the L1 and L2 caches can quickly pass data back and forth until ultimately, the L1 cache passes the data to the core.
  • L3 cache is shared by all cores in a CPU, and it acts as a buffer for passing data between the CPU and main memory. Sizes vary widely. In an 8-core AMD EPYC processor, the L3 cache is just 64MB. But in AMD’s 192-core CPU, the L3 Cache gets as big as 348MB.

Some AMD CPUs, including the AMD EPYC 9845, also include a 3D V-Cache. This AMD innovation stacks an additional cache on top of the L3 cache (hence the name 3D). Stacking the two caches vertically adds storage without increasing the overall size of the CPU.

The added 3D V-Cache also improves performance for workloads that benefit from a larger cache. Examples include scientific simulations and big data.

Nodes: Power & Efficiency

When it comes to CPU nodes, smaller is better. A smaller node size can deliver benefits that include lower power consumption, increased efficiency, and more compute performance per watt.

Nodes are expressed in nanometers (nm)—that’s one billionth of a meter—which describe the tiny size of transistors on a chip.

The latest AMD EPYC 9005-series architectures, ‘Zen 5’ and ‘Zen 5c,’ are built on 4nm and 3nm nodes, respectively.

Each of these individual performance gains may seem tiny when considered on a per-chip basis. But in the aggregate, they can make a huge difference. That’s especially true for resource-intensive workloads such as AI training and inferencing.

Coming Soon: Smaller, Faster CPUs

AMD’s near-term roadmap tells us we can expect its AMD EPYC CPUs to continue getting smaller, faster and more efficient.

Those manufacturing and performance gains will likely come from more cores per CPU socket, bigger and more efficient caches. Earlier this year, AMD said the next generation of its EPYC processors, codenamed Venice, will be brought up on TSMC’s advanced 2nm process technology.

Enterprises will be able to parlay those improvements into better performance under multi-tenant loads and reduced latency overall. The latter is particularly vital for modern operations.

The bottom lie: Denser CPU cores mean big business, both for processor makers such as AMD and the server vendors such as Supermicro that rely on these CPUs.

Denser CPUs are also vital for enterprises now transforming their data centers for AI. Because adding space is so slow and costly, these organizations are instead looking to pack more compute power per rack. Smaller, more powerful CPUs are an important part of their solution.

Minimum CPU size with maximum power? It’s coming soon to a data center near you.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Supermicro adds MicroBlade for CSPs powered by AMD EPYC 4005 series processors

Featured content

Supermicro adds MicroBlade for CSPs powered by AMD EPYC 4005 series processors

To serve cloud service providers, Supermicro adds a 6U, 20-node MicroBlade server powered by AMD EPYC 4005 series processors.

Learn More about this topic
  • Applications:
  • Featured Technologies:

Not every cloud service provider is as big or deep-pocketed as the big three—AWS, Google and Microsoft. And to serve those smaller and midsize CSPs, Supermicro recently added a 6U, 20-node server to its MicroBlade family powered by AMD EPYC 4005 series processors.

Smaller CSPs represent a big market. To be sure,  AWS, Google and Microsoft collectively drew nearly 65% of total worldwide cloud infrastructure services revenue in this year’s third quarter, according to Synergy Research Group.

But for both smaller CSPs and their suppliers, that remaining 35% was still quite valuable. Synergy estimates worldwide cloud infrastructure services revenue in Q3 totaled $107 billion. That means the share left to smaller and midsize CSPs was about $37 billion.

MicroBlade, Macro Benes

To serve these smaller CSPs, Supermicro recently introduced a 6U, 20-node MicroBlade (model number MBA-315R-1G) powered by a single AMD EPYC 4005 series processor.

This MicroBlade system delivers a cost-effective, green computing solution. It’s intended for workloads that include not only cloud computing, but also web hosting, dedicated hosting, virtual desktop infrastructure (VDI), AI inferencing, and enterprise workloads.

Supermicro CEO Charles Liang calls the new servers “a very cost-effective, green computing solution for cloud service providers.”

Key benefits of the new Supermicro system include up to 95% cable reduction with two integrated Ethernet switches per server; 70% space savings; and 30% energy savings over traditional 1U servers.

The system offers 3.3x higher density than a traditional 1U server. As a result, users can pack as many as 160 servers with 2,650 CPU cores, as well as 16 Ethernet switches, in a single 48U rack.

Under the hood, each MicroBlade server blade supports a single AMD EPYC 4005 CPU with up to 16 cores and 192GB of DDR5 memory. Also supported is a dual-slot, full-height/full-length (FHFL) GPU.

Also, this Supermicro system contains a dual-port 10GbE network switch. It’s designed to simplify topologies and enable more server instances per rack.

The 6U MicroBlade chassis can hold up to 20 individual server blades, two Ethernet switches and two management modules.

To protect workloads such as dedicated hosting, VDI, online gaming and AI inferencing, the Supermicro system also offers N+N redundancy. This setup configures two sets of independent components to provide high levels of reliability.

The MicroBlade system will also be available as a motherboard (model number BH4SRG) for Supermicro A+ servers.

 

 

Inside the AMD EPYC 4005 Series

The processors powering the new Supermicro server, AMD’s EPYC 4005 series, offer powerful performance for AI, cloud and hosting workloads. Yet they’re attractively priced for smaller businesses and hosting services.

The processors are based on the same core generation, ‘Zen 5,’ as are AMD’s more powerful data center processors, the AMD EPYC 9005 series. Yet the 4005 series processors have been designed for smaller operations, offering a combination of affordability, efficiency and ease of use.

AMD’s corporate VP for enterprise and HPC, Derek Dicker, says the AMD EPYC 4005 series processors “give our technology partners the flexibility to create powerful yet affordable systems that meet the specific needs of growing businesses and dedicated hosters.”

Do you have CSP clients looking for an affordable yet powerful servers? Tell them about these new AMD-powered Supermicro servers, coming soon.

Do More:

 

Featured videos


Events


Find AMD & Supermicro Elsewhere

Pages