Generative AI comes to Amazon Web Services
On Thursday, Amazon released a new suite of AI technologies, including foundational large language models (LLMs) called Titan and a cloud computing service called Bedrock, reports Reuters. The move comes as competitors Microsoft and Google integrate AI chatbots into their search engines and cloud operations.
Recently, LLMs like OpenAI’s GPT-4 and its instruction-tuned cousin ChatGPT have become one of the hottest stories in tech, inspiring large investments in AI labs and shaking up business operations in major players. LLMs, often grouped with similar technologies under the umbrella term “generative AI,” can take any kind of written input and transform, translate, or interpret it in different ways.
In response, Amazon’s cloud computing division, Amazon Web Services (AWS), has designed its new AI technologies to help companies develop their own chatbots and image-generation services (such as OpenAI’s DALL-E). As with Microsoft’s Azure cloud platform that powers OpenAI’s models, Amazon stands to reap financial rewards from renting out the computing muscle to make its own brand of generative AI happen.
AWS’s core offering, named Bedrock, enables businesses to customize “foundation models” using their own data. In AI, foundation models are core AI technologies that serve as starting points for companies to build upon. Using fine-tuning (additional training with a specific goal), companies can incorporate proprietary data, customizing the models for their own needs. OpenAI offers a similar service that allows customers to fine-tune models for custom chatbots.
Bedrock will provide customers with Amazon’s proprietary LLM foundation models, known collectively as Amazon Titan, as well as models from third-party companies. Startups AI21 Labs, Anthropic, and Stability AI will offer their models alongside Amazon’s.
According to Amazon, one key aspect of providing its Bedrock service will be allowing AWS customers to test these new AI technologies without managing the underlying data center servers that power them. Those data centers are costly investments. Amazon says the underlying servers for the Bedrock service will utilize a combination of Amazon’s custom AI chips (AWS Trainium and AWS Inferentia) and GPUs from Nvidia, which is currently the largest supplier of AI chips.
Also on Thursday, Amazon announced a preview of Amazon CodeWhisperer, an AI-powered coding assistant similar to GitHub Copilot and Replit Ghostwriter. It’s free for individual use and available for evaluation today.
https://arstechnica.com/?p=1931614