Amazon Web Services added capacity sharing and training plans to Amazon SageMaker Hyperpod and added Lumi AI and Poolside models to Amazon Bedrock's selection of third party models. AWS also launched Amazon Bedrock Marketplace.
The news was announced at a re:Invent 2024 keynote by Dr. Swami Sivasubramanian, VP of AI and Data at AWS. AWS add-on to the previous updates to SageMaker and Bedrock.
- AWS unveils next-gen Amazon SageMaker in bid to unify data, analytics, AI
- AWS aims to make Amazon Bedrock your agentic AI point guard
With Amazon SageMaker HyperPod, AWS added the following:
- Capacity sharing and allocation and governance tools so enterprises to share large pools of compute, set priorities for teams, projects and tasks and schedule them based on priorities.
- Training plans so customers can optimize genAI models based on hardware, budget, timelines and region constraints. Training plans will automatically move work across availability zones.
- Recipes for data scientists and engineers to start training and fine-tuning popular foundation models in hours. These recipes are curated for training and ready to use for popular models. These recipes can be set up to swap hardware to optimize performance and lower costs.
For Bedrock, AWS added new models from Luma AI, a specialist in creating video clips from text and images, and Poolside, which specializes in models for software engineering. Amazon Bedrock has also expanded models from its current providers such as AI21 Labs, Anthropic, Cohere, Meta, Mistral, Stability.ai and Amazon.
The addition of Amazon Bedrock Marketplace will give customers the ability to try models and balance cost and performance. Amazon Bedrock Marketplace has access to more than 100 emerging and specialized foundation models.
Sivasubramanian said:
"With Bedrock, we are committed to giving you access to the best model for all your use cases. However, while model choice is critical, it's really just the first step when we are building for inference. Developers also spend lot of time evaluating models for their needs, especially factors like cost and latency that require a delicate balance."
Bedrock also was updated with intelligent prompt routing, which will automatically route requests among foundation models in the same family. The aim is to provide high-quality responses with low cost and latency. The routing will be based on the predicted performance of each request. Customers can also provide ground truth data to improve predictions.
More from re:Invent:
- Amazon Q Business gets a story at AWS re:Invent 2024
- PagerDuty integrates with Amazon Bedrock, Q Business: Will it boost large enterprise traction?
- AWS launches Amazon Nova foundation models in commoditization play
- AWS aims to make Amazon Bedrock your agentic AI point guard
- AWS revamps S3, databases with eye on AI, analytics workloads
- AWS scales up Trainium2 with UltraServer, touts Apple, Anthropic as customers
- AWS outlines new data center, server, cooling designs for AI workloads
- AWS re:Invent 2024: Four AWS customer vignettes with Merck, Capital One, Sprinkr, Goldman Sachs
- Oracle Database@AWS hits limited preview