How Amazon rebuilt Alexa with generative AI

Today, we introduced Alexa+, the next generation of Alexa. Rebuilt from the ground up, this new version of Alexa uses a state-of-the-art architecture that automatically connects a variety of large language models (LLMs), agentic capabilities, services, and devices at scale. This makes Alexa+ much more conversational, smarter, personalized, and capable of getting more things done for customers.

With an undertaking of this scale, the team had to solve many technical challenges along the way. Here are five of the biggest advances we made to unlock Alexa+ and deliver our next generation assistant, powered by generative AI.

We built an all-new architecture to connect to tens of thousands of services and devices

LLMs are great for conversations, but they don’t inherently support APIs, which are the core protocols to getting things done outside of a chat window and in the real world for customers—things like finding and booking an appointment or ordering your groceries. To augment the native capabilities of LLMs, we built an all-new architecture to orchestrate APIs at scale.

This architecture is what will let customers quickly and seamlessly connect with services they already use in their daily life: GrubHub, OpenTable, Ticketmaster, Yelp, Thumbtack, Vagaro, Fodor’s, Tripadvisor, Amazon, Whole Foods Market, Uber, Spotify, Apple Music, Pandora, Netflix, Disney+, Hulu, Max, smart home devices from companies like Philips Hue and Roborock, and so much more.

Curated: https://Layer3.host