Amazon says its next-gen chips are four times faster for AI training

Amazon Web Services (AWS) just kicked off its Las Vegas-based re:Invent conference with a stream of announcements, most of which involve the year’s most popular technology, AI. These news items, taken as a whole, give us a sneak peek at the company’s long-term goals for artificial intelligence platforms.

First of all, AWS unveiled its latest-generation of AI chips intended for model training and for running trained models. Trainium2, which is obviously for model training, has been designed to deliver up to 4x better performance and 2x energy efficiency when compared to its forebear. Amazon promises these chips will allow programmers to train models quickly and at a lower cost, due to a reduction in energy use. Anthropic, an Amazon-backed OpenAI competitor, has already announced plans to build models using Trainium2 chips.

Graviton4, on the other hand, is more for general use. These processors are based on Arm architecture, but consume less energy than Intel or AMD chips. Amazon promises an increase of 30 percent in general performance when using a trained AI model embedded within a Graviton4 processor. This should lower cloud-computing costs for organizations that regularly employ AI models and offer a slight uptick in speed for regular users just looking to make some fake photos of Harry Potter at a rave or whatever.

All told, Graviton4 should allow AWS customers to “process larger amounts of data, scale their workloads, improve time-to-results and lower their total cost of ownership.” It’s available today in preview with a wider release planned for the coming months.

Typically, when a company announces new in-house chips, that spells trouble for current third-party providers like NVIDIA. The company is a huge player in the enterprise AI space, thanks to companies using its GPUs for training and its Arm-based datacenter CPU Grace. Instead of eschewing the partnership in favor of proprietary chips, Amazon is further cementing the relationship by offering enterprise customers cloud access to NVIDIA’s latest H200 AI GPUs. It’ll also operate more than 16,000 Nvidia GH200 Grace Hopper Superchips expressly for NVIDIA’s research and development team. This is a similar approach to its chief AI rival, Microsoft, which also announced an enhanced partnership with NVIDIA at the same time it revealed its proprietary AI chip, Maia 100.

Amazon also announced a new business-focused AI chatbot called Q, a name that was likely inspired by the Star Trek demigod and not the Trump-adjacent conspiracy peddler. It’s described as a “new type of generative AI-powered personal assistant” and is specifically designed to help streamline work projects and customer service tasks. It can be tailored to suit any business and offers relevant answers to commonly-asked questions. Amazon Q can also generate content on its own and take actions based on customer requests. It’ll even customize interactions based on a user’s role within a company.

It’ll exist on communication apps like Slack and in text-editing applications commonly-used by software developers. To that end, Q can actually change source code and can connect to more than 40 enterprise systems, including Microsoft 365, Dropbox, Salesforce and Zendesk, among others. Amazon Q is currently available in preview, with a wider release coming soon. It’ll cost anywhere from $20 to $30 per user each month, depending on available features.

So what have we learned here? Amazon is betting big on AI, like everyone else. More specifically, it’s battling with old cloud rival Microsoft to be the go-to company for enterprise-based AI. It’s also using AI to continue its dominance in the cloud computing space, hoping to minimize any increase in market share for Microsoft and other players like Google and Alibaba.

This article originally appeared on Engadget at https://www.engadget.com/amazon-says-its-next-gen-chips-are-four-times-faster-for-ai-training-200028416.html?src=rss https://www.engadget.com https://www.engadget.com/amazon-says-its-next-gen-chips-are-four-times-faster-for-ai-training-200028416.html?src=rss
созданный 2y | 28 нояб. 2023 г., 21:30:14


Войдите, чтобы добавить комментарий

Другие сообщения в этой группе

Anthropic brings web search to free Claude users

Anthropic is continuing to trickle down features to its free users. The latest one to make the leap out of subscriber-only mode is web search, which the company

28 мая 2025 г., 00:40:11 | Engadget
Apple buys the maker of Sneaky Sasquatch

Apple has bought a game studio for the first time. Digital Trends

27 мая 2025 г., 22:20:13 | Engadget
Texas enacts age-verification law for app stores

Texas is the latest state to adopt an age-verification law for app stores. Despite lobbying from big tech, reportedly including a personal

27 мая 2025 г., 20:10:14 | Engadget
Video Games Weekly: Grand Theft Auto is no friend to the queer community

Welcome to the initial installment of Video Games Weekly on Engadget. Expect a new story every Monday (yes, we realize today is Tuesday), broken into two parts. The first is a space for short e

27 мая 2025 г., 20:10:13 | Engadget
EU regulators are investigating Pornhub and three other sites

European regulators are investigating Pornhub. On Monday, the European Commission (EC) accused

27 мая 2025 г., 20:10:12 | Engadget
Cities: Skylines II delays Bridges & Ports DLC to Q4

The saga of Cities: Skylines II has involved a lot of frustration, and the years-long ride still doesn't appear to be over. The latest news from the developer is yet another delay, this ti

27 мая 2025 г., 20:10:11 | Engadget
The Browser Company stops active development of Arc in favor of new AI-focused product

The Browser Company has stopped active development of the popular Arc web browser, according

27 мая 2025 г., 17:40:19 | Engadget