Off-Prem

GenAI will be bigger than the cloud or the internet, Amazon CEO hopes

And Andy Jassy will happily take your money along the way


It's safe to say Amazon CEO Andy Jassy is pretty jazzed about generative AI's potential to drive profits.

"Generative AI may be the largest technology transformation since the cloud — which itself, is still in the early stages — and perhaps since the internet," he wrote in his annual letter to shareholders Thursday. "There never has been a time in Amazon's history where we've felt so much opportunity to make our customer's lives better and easier."

For the e-commerce giant turned cloud titan, the chances to capitalize on the much-hyped technology are many, but just like the move to the cloud, there is plenty of work that needs done first.

The stage, what Jassy describes as the bottom layer, won't come as a surprise to anyone: selling developers the computing resources and software services necessary to develop their own foundation models.

These efforts, he acknowledges, will require a lot of Nvidia GPUs. Even if Amazon can't build the shovels — or, at least not the ones everyone wants — in this AI goldrush, it can still rent them at a massive premium.

At GTC, Nvidia revealed that Amazon's Ceiba supercomputer, which was initially planned to use its existing Hopper GPUs, will actually be built using more than 20,000 Blackwell accelerators.

But, with those resources in high-demand, Jassy says there's growing interest in alternative hardware, highlighting Anthropic's plans to use Amazon's own home-grown Trainium and Inferentia chips to build, train and deploy future foundation models. On the software side, Jassy also highlighted the cloud-giants SageMaker platform which is designed to streamline the process of training said models.

To support these efforts Amazon also plans to expand its datacenter presence with six new regions in Malaysia, Mexico, New Zealand, Saudi Arabia, Thailand, and Germany already in the works.

One step up from this, Jassy says Amazon is well positioned to help enterprises fine-tune, customize, and integrate existing foundation models — something we'll note is computationally much easier to do than train a system from scratch.

 For this Amazon has Bedrock, which it announced in early 2023. According to Jassy, the service has added tens of thousands of users in the year since its launch. It probably doesn't hurt that Amazon is handing out $500k in credits to anyone who wants to play with pre-trained models.

The idea of customizing and building applications using one or more large language models (LLMs) has garnered a lot of attention as of late. During his keynote address this week at Intel Vision, CEO Pat Gelsinger discussed using technologies like retrieval augmented generation (RAG) to make general purpose models, like Llama2, context aware by connecting them to vector databases containing relevant information.

Nvidia CEO Jensen Huang has also sung RAG's praises as one of the key technologies required to bring generative AI apps to the enterprise.

Amazon, for its part, has partnered closely with AI startups, including a $4 billion investment in Anthropic, the model builder behind Claude.

At the top of Jassy's generative AI stack are off-the-shelf applications and services. These are apps like Amazon's Q code assistant, which debuted last fall, as well as internal applications designed to drive sales.

"We're building a substantial number of GenAI applications across every Amazon customer business. These range from Rufus — our new AI-powered shopping assistant — to an even more intelligent and capable Alexa, to advertising capabilities," he said.

While Amazon will continue to develop these kinds of AI tools, Jassy expects most generative AI apps will ultimately be built by others.

"These AWS services, at all three layers of the stack, comprise a set of primitives that democratize this next seminal phase of AI, and will empower internal and external builders to transform virtually every customer experience that we know," he said. ®

Send us news
15 Comments

Amazon congratulates itself for AI code that mostly works

Web services souk celebrates 'leader' designation for Q Developer

Canadian artist wants Anthropic AI lawsuit corrected

Tim Boucher objects to the mischaracterization of his work in authors' copyright claim

All that new AI-fueled datacenter space? Yeah, that's mostly ours – cloud giants

Construction surged 70% ... with 80% already snapped up

AI firms propose 'personhood credentials' … to fight AI

It's going to take more than CAPTCHA to prove you're real

AMD's Victor Peng: AI thirst for power underscores the need for efficient silicon

Moore's Law may be running out of steam, but there are still knobs to turn and levers to pull

AI's thirst for water is alarming, but may solve itself

Its energy addiction, on the other hand, only seems to get worse

'Error' causes Alexa to endorse Kamala Harris, refuse to discuss Trump

Bot shouldn't have political opinions, says Amazon

DataVita declares sovereignty with 'National Cloud' for UK

Scottish provider promises no hidden fees, full control, and safe haven for data

Tenstorrent's Blackhole chips boast 768 RISC-V cores and almost as many FLOPS

Shove 32 of 'em in a box and you've got nearly 24 petaFLOPS of FP8 perf

India misses ten million public WiFi hotspot goal, catchup not looking likely

Plus: Glowing reports for Fukushima wastewater; New datacenters in Fiji, Malaysia & South Korea; and more

The future of AI/ML depends on the reality of today – and it's not pretty

The return of Windows Recall is more than a bad flashback

China AI devs use cloud services to game US chip sanctions

Orgs are accessing restricted tech, raising concerns about more potential loopholes