Deals Counterparts

AWS partners with Cerebras Systems for AI inference solution deployment on Amazon Bedrock in AWS data centers

Partnership Data Center announced Mar 13, 2026
deployment
Stage

Amazon Web Services (AWS) has partnered with Cerebras Systems to deliver an AI inference solution supporting generative AI applications and LLM workloads. The solution will combine AWS Trainium-powered servers with Cerebras' wafer-scale CS-3 systems and Elastic Fabric Adapter (EFA) networking, deployed on Amazon Bedrock in AWS data centers. This partnership aims to improve AI inference speed and performance by disaggregating workloads across Trainium and CS-3 hardware.

Get daily data center deal alerts — free, no spam.

Subscribe to access full counterpart details, deal analysis, and timeline.

Subscribe from €20/mo →

Deal Analysis

AWS, the world's leading cloud platform, is integrating Cerebras Systems' specialized AI hardware directly into its data centers. Announced for March 2026, this partnership will deploy Cerebras' wafer-scale CS-3 systems alongside AWS's own Trainium-powered servers on Amazon Bedrock. The aim is to disaggregate generative AI and LLM workloads, leveraging Elastic Fabric Adapter (EFA) networking to combine the strengths of both hardware types. This move by AWS deepens its core infrastructure investment, seeking to boost AI inference speed and performance for demanding applications.
  • AWS Trainium servers paired with Cerebras wafer-scale CS-3 systems.
  • Deployment on Amazon Bedrock within AWS data centers.
  • Partnership announced for March 13, 2026.
  • Aims to disaggregate generative AI workloads for performance gains.

Source Intelligence

KEY DETAILS

The solution will combine AWS Trainium-powered servers with Cerebras' wafer-scale CS-3 systems and Elastic Fabric Adapter (EFA) networking. AWS will also offer open-source LLMs and Amazon Nova, its own foundation models, using Cerebras hardware “later this year.” The combined Trainium/CS-3 solution will enable “inference disaggregation,” splitting AI inference into a compute intensive prompt processing (‘prefill’) stage and a memory bandwidth-intensive output generation (‘decode’) stage. Trainium chips will be optimized for prefill, while Cerebras CS-3 hardware will be optimized for decode.

Deal Size
The financial terms of the agreement have not been disclosed.
Location
set to be deployed on Amazon Bedrock in AWS data centers.
Financials
The financial terms of the agreement have not been disclosed.
Announcement
March 13, 2026 By Charlotte Trueman
PARTIES MENTIONED IN SOURCE
A
Amazon Web Services (AWS) partner

"Amazon Web Services (AWS) has partnered with Cerebras Systems"

C
Cerebras Systems partner

"Amazon Web Services (AWS) has partnered with Cerebras Systems"

medium quality Enriched Mar 13, 2026

Timeline

Announced
Mar 13, 2026
Signed
Closed

Get the full picture — timeline, source intelligence, and counterpart analysis.

Subscribe from €20/mo →
Track Data Center deals