UPDATED 13:50 EDT / APRIL 13 2023

AI

AWS doubles down on generative AI training

Amazon Web Services Inc. is extending its reach further into the domain of artificial intelligence software development with the release today of several new tools for generative AI  training and deployment on its cloud platform.

In a post on the AWS Machine Learning blog, the company detailed new offerings that include the ability to build and train foundation models, which are large-scale, pre-trained language models that create a foundation for targeted natural language processing tasks. Foundation models are typically trained on massive amounts of text data using deep learning techniques, which allows them to learn to understand the nuances of human language to the point that they can generate text that is almost indistinguishable from that written by humans.

The use of pre-trained foundation models can save developers significant amounts of time and resources that would otherwise be required to train a language model from scratch. OpenAI LLC’s Generative Pre-trained Transformer or GPT is an example of a foundation model that can be used for text generation, sentiment analysis and language translation.

LLM choices

Bedrock is a new service that makes foundation models from a variety of sources available via an application program interface. They include the Jurassic-2 multilingual large language models from AI21 Labs Ltd. — which generate text in Spanish, French, German, Portuguese, Italian and Dutch — and Anthropic’s PBC’s Claude LLM, which performs a variety of conversational and text processing tasks based on responsible AI system training principles. Users can also access Stability AI Ltd. as well as Amazon LLMs using the API.

Foundation models are pre-trained at internet scale and so can be customized with relatively little additional training, wrote Swami Sivasubramanian, vice president of database, analytics and machine learning at AWS. He gave the example of a content marketing manager for a fashion retailer who can provide Bedrock with as little as 20 examples of well-performing taglines “from past campaigns, along with the associated product descriptions, and Bedrock will automatically start generating effective social media, display ad and web copy for the new handbags.”

In conjunction with the Bedrock announcement, AWS is also rolling out two new large language models under the Titan banner. The first is a generative LLM for summarization, text generation, classification, open-ended question-and-answer and information extraction. The second is an LLM that translates text inputs into numerical representations that contain the semantic meaning of the text and are useful in producing contextual responses that go beyond word matching.

Noticeably absent from the announcement was any mention of OpenAI, in whom Microsoft Corp. is a major investor, but that shouldn’t be an impediment for Amazon given the market’s appetite for large language models.

“There is a rush to create many of them,” said Rajesh Kandaswamy, a distinguished analyst and fellow at Gartner Inc. “Pretty much any technology you’re going to see at this stage will have options from multiple innovators.”

AWS trails Microsoft and Google LLC in bringing its own large LLM to market, but that shouldn’t be seen as a competitive handicap, Kandaswamy said. “I don’t think anyone is so behind that they have to play catchup,” he said. “It might appear that there is a big race, but the customers we speak with, other than very early adopters, have no idea what to do with it.”

Hardware boost

AWS is also beefing up the hardware it uses to deliver training and inferencing on its cloud. New, network-optimized EC2 Trn1n instances, which incorporate the company’s proprietary Trainium and Inferentia2 processors, now provide 1,600 gigabits per second of network bandwidth, or roughly a 20% performance boost. The company’s Inf2 instances, which use Inferentia2 for inferencing of large-scale generative AI applications with models containing hundreds of billions of parameters, are also now generally available.

Another availability announcement is CodeWhisperer, an AI coding companion that uses a foundation model to generate code suggestions in real time based on natural language comments and prior code in an integrated development environment. The tool works with Python, Java, JavaScript, TypeScript C# and 10 other languages and can be accessed from a variety of IDEs.

“Developers can simply tell CodeWhisperer to do a task, such as ‘parse a CSV string of songs’ and ask it to return a structured list based on values such as artist, title and highest chart rank,”  Sivasubramanian wrote. CodeWhisperer generates “an entire function that parses the string and returns the list as specified.” He said developers who used the preview version reported speed improvements of 57% and a 27% better success rate than when working without the tool.

The LLM picture is likely to remain fragmented and chaotic for the immediate future as many players try to cash in on the success of proofs of concept like ChatGPT. It’s unlikely that any one model will come to dominate the market as Google’s Natural Language API has in speech recognition, Kandaswamy said.

“Just because a model is good at one thing doesn’t mean it’s going to be good with everything,” he said. “It’s possible over two or three years everybody will offer everybody else’s model. There will be more blending and cross-technology relationships.”

Image: Unsplash

A message from John Furrier, co-founder of SiliconANGLE:

Your vote of support is important to us and it helps us keep the content FREE.

One click below supports our mission to provide free, deep, and relevant content.  

Join our community on YouTube

Join the community that includes more than 15,000 #CubeAlumni experts, including Amazon.com CEO Andy Jassy, Dell Technologies founder and CEO Michael Dell, Intel CEO Pat Gelsinger, and many more luminaries and experts.

“TheCUBE is an important partner to the industry. You guys really are a part of our events and we really appreciate you coming and I know people appreciate the content you create as well” – Andy Jassy

THANK YOU