SEATTLE & SUNNYVALE, Calif.–(BUSINESS WIRE)–Amazon Web Services, Inc. (AWS), an Amazon.com, Inc. company (NASDAQ: AMZN), and Cerebras Systems today announced a collaboration that will, in the coming months, deliver the fastest AI inference solutions available for generative AI applications and LLM workloads. The solution, to be deployed on Amazon Bedrock in AWS data centers, combines AWS Trainium-powered servers, Cerebras CS-3 systems, and Elastic Fabric Adapter (EFA) networking. Later this year, AWS will also offer leading open-source LLMs and Amazon Nova using Cerebras hardware.
“Partnering with AWS to build a disaggregated inference solution will bring the fastest inference to a global customer base. Every enterprise will benefit from blistering fast inference within their existing AWS environment” – Andrew Feldman, Cerebras CEO
“Inference is where AI delivers real value to customers, but speed remains a critical bottleneck for demanding workloads like real-time…