On the outskirts of Austin, Texas, lies Annapurna Labs—a combination of offices, workshops, and even a mini “data center”—where Amazon Web Services (AWS) engineers are designing the future of computing.
High-tech electronics lab with workstations and testing equipment
Annapurna Labs is a specialist microelectronics company Amazon acquired in 2015. Like its parent company, Annapurna Labs got its start in a household garage, and a lot has changed but the scrappy startup mentality remains. The lab’s specialist engineers, computer scientists, operations and logistics experts, and more—take a hands-on approach to every aspect of developing custom silicon chips and accelerators. No job is too small, even the ones requiring microscopic probes.
An engineer in an AWS lab examines data visualizations, zooming in and out of elements on a large monitor display
We’re giving you a look behind the scenes at Annapurna Labs’ operations, where the team embraces the “organized chaos” of innovation.

Page overview

Designing with the whole system in mind

1
Designing with the whole system in mind
2
Iterating quickly, failing early, and fixing things fast
3
Co-designing software and hardware
4
Testing to ensure peak performance
5
Building the world’s most powerful AI computing systems
6
Foreseeing the future of computing
Designing with the whole system in mind
Team collaborating in modern office meeting room with whiteboard

The Annapurna Labs office looks like a typical working space with a mix of employees typing at desktops and brainstorming in conference rooms. Many of these employees are at the front line of machine learning acceleration, developing the layers of software that power silicon chips. They make up a critical part of Annapurna’s secret sauce—a system-first mindset.

“Instead of building a chip and then integrating it into a system and writing software to it, we flipped the process on its head,” said Ron Diamant, lead architect. “We first designed the full system and work backwards from that in order to specify the most optimal chip for that system. And this allows us to create a much more tailored chip for the workloads that we’re trying to accelerate.”

Iterating quickly, failing early, and fixing things fast

Past the rows of cubicles are three different labs where the hardware comes into play. Engineers at cable-covered workstations use power tools to build boards and specialized microscopes to view tiny chip components. Dozens of fans keep the equipment (and humans) cool as servers run in the background. Despite its scattered state, everything in the lab has a purpose and serves as a reminder that learning can be messy.

A Software Engineer in an AWS lab working on a computer with a wiring board displayed behind her

“When you go into the lab and you see equipment everywhere. It’s organized chaos,” said Rami Sinno, director of silicon engineering. “We iterate quickly, we fail early, and we fix it. And this is what allows us to continuously deliver very high-performance, low-cost products to our customers.”

Co-designing software and hardware
Two engineers examining computer components on workbench

Annapurna Labs’ vertically integrated process enables control of the entire stack of components required for machine learning accelerator servers. Both software and hardware engineers collaborate at every stage of development, from chip design to server deployment in AWS data centers.

“As we’re developing the chip, we develop the software in parallel. We use both of them in testing so that we make sure everything’s working well together and we can do trade-off analysis,” said Laura Sharpless, software engineering manager. “Every day I come into the office, I get to solve a new problem. Maybe today we’re working on hardware, physical boards. And tomorrow we’re looking at how do we actually scale the software and support multiple generations really seamlessly to scale faster.”

Testing to ensure peak performance
Microscope examining circuit board in electronics lab

Annapurna's testing and validation processes are critical to ensuring the reliability and robustness of components for 24/7 operation in AWS data centers. Engineers test all software and hardware components at every level from chip to board to server. The lab consists of stations where engineers use specialized equipment to introduce different variables like functionality, voltage, and temperature.

Display screen showing AWS chip performance testing process

“Testing significantly cuts down the development time so our software engineers can iterate faster,” said Prashant Pappu, principal hardware engineer, “and hardware engineers can focus on finding issues early on in the cycle.”

Building the world’s most powerful AI computing systems
Close-up of microchip repair with precision tool

Prior to the acquisition, Annapurna Labs and AWS worked together on the production of next-generation hardware AWS Nitro and its supporting hypervisor. Just over a decade later, Nitro is essential to every AWS server. The technology is the foundation of EC2 instances, enables AWS to innovate faster, further reduce cost for customers, and deliver increased security. Shortly after joining AWS, Annapurna Labs embarked on Graviton, its second product line. Now in its fourth generation, Graviton gives customers more computing capabilities while reducing their carbon footprint.

Hands holding small electronic device with digital display

Annapurna Labs’ machine learning chips—Inferentia and Trainium—are the third product line. Their names are a direct reflection of their use cases. Customers use Inferentia to run machine learning inference at scale and Trainium to run large-scale training workloads like generative AI and computer vision. Trainium2, the second-generation chip, is an essential part of Annapurna Labs’ development of increasingly powerful AI computing systems like Trainium2 instances and UltraServers.

The Build on Trainium program enables expanded research and training opportunities for advanced AI processing, sparking innovation for the frontier models of the future.

“An UltraServer combines four Trainium2 servers and 64 Trainium2 chips into one server with very fast connections between them,” said Tobias Edler von Koch, principal software engineer. “As machine learning models become too large to be handled definitely by an individual chip, or even by an individual machine, you need to scale out and have multiple servers collaborate.”

Foreseeing the future of computing
Data center server room with colorful network cables and technician

Annapurna Labs’ stealthy setup and eager engineers makes it uniquely suited to meet the demand for continuous innovation in the rapid race for AI advancement. Along with the development of its next-generation chips, Annapurna is partnering with AI startup Anthropic to take on its most ambitious challenge yet: building Project Rainier expected to be the world’s largest supercomputer.

Colorful ethernet wires in a data center

“It's so exciting to be in this fast-moving environment, innovating on behalf of customers and working closely with customers to make sure that we are building the right things in the future,” said Gadi Hutt, director of product and customer engineering. "My prediction is the next celebration in 20 years will come much faster because we’re having so much fun.”

To learn more about how and why AWS develops its family of custom chips and accelerators, visit the AWS Silicon Innovation Day page.