Amazon.com Inc.

07/23/2024 | News release | Distributed by Public on 07/23/2024 10:25

Llama 3.1 models from Meta are now available on AWS, offering more options for building generative AI applications

Meta's most advanced large language models (LLMs) give customers more choices when building, deploying, and scaling generative AI applications.
The next generation of Llama modelsfrom technology company Meta are available today on Amazon Web Services (AWS) via Amazon Bedrockand Amazon SageMaker. They are also available via Amazon Elastic Cloud Compute (Amazon EC2) using AWS Trainium and Inferentia.
The Llama 3.1 models are a collection of pretrained and instruction fine-tuned large language models (LLMs)in 8B, 70B, and 405B sizes that support a broad range of use cases. They are particularly suited for developers, researchers, and businesses to use for text summarization and classification, sentiment analysis, language translation, and code generation.

AWS debuts 2 AI certifications to give you an edge in pursuing in-demand cloud jobs

People with AI skills can boost their salaries by 47%, and AWS is launching a suite of new certifications and courses to help grow those skills.

Read more
According to Meta, Llama 3.1 405B is one of the best and largest publicly available foundation models (FMs), setting a new standard for generative AI capabilities. It is particularly well suited for synthetic data generation and model distillation, which improves smaller Llama models in post-training. The models also provide state-of-the-art capabilities in general knowledge, math, tool use, and multilingual translation.
All of the new Llama 3.1 models demonstrate significant improvements over previous versions, thanks to vastly increased training data and scale. The models support a 128K context length, an increase of 120K tokens from Llama 3. This means 16 times the capacity of Llama 3 models and improved reasoning for multilingual dialogue use cases in eight languages: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
Other improvements include an enhanced grasp of linguistic nuances, meaning Llama 3.1 shows improved contextual understanding and can handle complexities more effectively. The models can also access more information from lengthy passages of text to make more informed decisions, as well as leverage richer contextual data to generate more subtle and refined responses.
"We have a longstanding relationship with Meta and are excited to make their most advanced models available to AWS customers today," said Matt Garman, CEO of AWS. "Customers love the ability to customize and optimize Llama models specific to their individual use cases, and with Llama 3.1 now available on AWS, customers have access to the latest state-of-the-art models for building AI applications responsibly."

AWS's head of security shares 7 reasons why security will always be Amazon's top priority

AWS executive Chris Betz discusses how security is deeply embedded into Amazon's culture and shares tools to help protect customers in the era of generative AI.

Read more
For the past decade, Meta has been focused on putting tools into the hands of developers, and fostering collaboration and advancements among developers, researchers, and organizations. Llama models are available in a range of parameter sizes, enabling developers to select the model that best fits their needs and inference budget. Llama models on AWS open up a world of possibilities because developers don't need to worry about scalability or managing infrastructure. AWS offers a simple turnkey way to get started using Llama.
"Open source is the key to bringing the benefits of AI to everyone," said Mark Zuckerberg, founder and CEO, Meta. "We've been working with AWS to integrate the entire Llama 3.1 collection of models into Amazon SageMaker JumpStart and Amazon Bedrock, so developers can use AWS's comprehensive capabilities to build awesome things like sophisticated agents that can tackle complex tasks."

Benefits of the new Llama 3.1 models

Llama 3.1 405B

  • Ideal for enterprise applications and research and development (R&D)
  • Use cases include: long-form text generation, multilingual and machine translation, coding, tool use, enhanced contextual understanding, and advanced reasoning and decision-making

Llama 3.1 70B

  • Ideal for content creation, conversational AI, language understanding, and R&D
  • Use cases include: text summarization, text classification, sentiment analysis and nuanced reasoning, language modeling, code generation, and following instructions

Llama 3.1 8B

  • Ideal for limited computational power and resources, as well as mobile devices
  • Faster training times
  • Use cases include: text summarization and classification, sentiment analysis, and language translation

AWS offers easy access to a wide range of LLMs

Amazon Bedrock, which offers tens of thousands of customers secure, easy access to the widest choice of high-performing, fully managed LLMs and other FMs, as well as leading ease-of-use capabilities, is the easiest place for customers to get started with Llama 3.1, with 8B and 70B generally available and 405B available in preview.
Customers seeking to access Llama 3.1 models and leverage all of AWS's security and features can easily do this in Amazon Bedrock with a simple API, and without having to manage any underlying infrastructure. At launch, customers can take advantage of the responsible AI capabilities provided by Llama 3.1. These models will work with Amazon Bedrock's data governance and evaluation features, including Guardrails for Amazon Bedrockand Model Evaluation on Amazon Bedrock. Customers will also be able to customize the models using fine-tuning, which is coming soon.

AWS announces 5 new innovations to help everyone build with generative AI

AI offerings for all skill levels, plus a free training milestone for cloud, and generative AI. Here's all the news announced from AWS Summit in New York.

Read more
Amazon SageMaker is the best place for data scientists and ML engineers to pre-train, evaluate, and fine-tune FMs with advanced techniques, and deploy FMs with fine-grain controls for generative AI use cases that have stringent requirements on accuracy, latency, and cost. Today, customers can discover and deploy all Llama 3.1 models in just a few clicks via Amazon SageMaker JumpStart. With fine-tuning coming soon, data scientists and ML engineers will be able to take building with Llama 3.1 one step further-for example, by adapting Llama 3.1 on their specific datasets in mere hours.
"Amazon Bedrock is the easiest place to quickly build with Llama 3.1, leveraging industry-leading privacy and data governance, evaluation features, and built-in safeguards. Amazon SageMaker-with its choice of tools and fine-grain control-empowers customers of all sizes and across all industries to easily train and tune Llama models to power generative AI innovation on AWS," Garman said.
For customers who want to deploy Llama 3.1 models on AWS through self-managed machine learning workflows for greater flexibility and control of underlying resources, Amazon Elastic Compute Cloud (Amazon EC2)accelerated computing provides a broad choice of compute options. AWS Trainium and AWS Inferentia2 enable high performance, cost-effective fine-tuning, and deployment for Llama 3.1 models on AWS. Customers can get started with Llama 3.1 on AWS AI chips using Amazon EC2 Trn1 and Inf2 instances.

Customers are already using Llama models on AWS

  • Nomura, a global financial services group spanning 30 countries and regions, is using Llama models in Amazon Bedrock to simplify the analysis of dense industry documents to extract relevant business information, empowering employees to focus more time on drawing insights and deriving key intel from data sources like log files, market commentary, or raw documents. Check out the work Nomura is doing with Llama.
  • TaskUs, a leading provider of outsourced digital services and customer experiences, uses Llama models in Amazon Bedrock to power TaskGPT, a proprietary generative AI platform, on which TaskUs builds intelligent tools that automate parts of the customer service process across channels, freeing up teammates to address more complex issues and delivering better customer experiences overall.
Get started with Llama on Amazon Bedrockand learn more on the AWS News blog.
For more information about Amazon SageMaker, visit the AWS Machine Learning blogand AWS website.
Amazon Bedrock and Amazon SageMaker are part of distinct layers of the "generative AI stack." Find out what this means, and why Amazon is investing deeply across all three layers.
To learn more about how AWS AI chips deliver high performance and low cost for Meta Llama 3.1 models on AWS, visit the AWS Machine Learning Blogand take a look inside the lab where AWS makes custom chips.