[ad_1]
During the AWS re:Invent generative AI keynote, Amazon introduced Bedrock help for Claude 2.1 and Llama 2 70B and extra.
After the AWS bulletins yesterday in regards to the Amazon Q chatbot for enterprise and highly effective new chips for AI workloads, Vice President of Databases, Analytics and Machine Learning at AWS Swami Sivasubramanian took the stage on the AWS re:Invent convention in Las Vegas on Nov. 29 to dive deeper into AWS AI choices. He introduced new generative AI fashions coming to Amazon Bedrock, multimodal looking obtainable for Amazon Titan in Amazon Bedrock and plenty of different new enterprise software program options and instruments associated to utilizing generative AI for work.
Jump to:
Amazon Titan can now run searches based mostly on textual content and pictures
Amazon Titan Multimodal embeddings are actually typically availability in Amazon Bedrock, the AWS instrument for constructing and scaling AI functions. Multimodal embeddings permit organizations to construct functions that allow customers search utilizing textual content and pictures for richer search and advice choices, mentioned Sivasubramanian.
“They (AWS customers) want to enable their customers to search for furniture using a phrase, image or even both,” mentioned Sivasubramanian. “They could use instructions like ‘show me what works well with my sofa’.”
SEE: Are AWS or Google Cloud proper for your online business? (TechRepublic)
Titan Text Lite and Titan Text Express added to Amazon Bedrock
Titan Text Lite and Titan Text Express are actually usually obtainable in Amazon Bedrock to assist optimize for accuracy, efficiency and price, relying on their use instances. Titan Text Lite is a really small mannequin for textual content and may be fine-tuned. Titan Text Express is a mannequin that may do a wider vary of text-based generative AI duties, corresponding to conversational chat and open-ended questions.
Titan Image Generator (Figure A) is now obtainable in public preview within the U.S. It can be utilized to create photographs utilizing pure language prompts. Organizations can customise photographs with proprietary information to match their trade and model. Images will probably be invisibly watermarked by default to assist keep away from disinformation.
Figure A

Claude 2.1 and Llama 2 70B now hosted on Amazon Bedrock
Amazon Bedrock will now help Anthropic’s Claude 2.1 for customers within the U.S. This model of the Claude generative AI affords developments in a 20,000 context window, improved accuracy, 50% fewer hallucinations even throughout adversarial immediate assaults and two occasions discount in false statements in open-ended conversations in comparison with Claude 2. Tool use for perform calling and workflow orchestration in Claude 2.1 can be found in beta for choose early entry companions.
Meta’s Llama 2 70B, a public giant language mannequin fine-tuned for chat-based use instances and large-scale duties, is offered as we speak in Amazon Bedrock.
Claude help obtainable in AWS Generative AI Innovation Center
The AWS Generative AI Innovation Center will increase early in 2024 with a customized mannequin program for Anthropic Claude. The AWS Generative AI Innovation Center is designed to assist individuals work with AWS’ group of specialists to customise Claude wants for one’s personal proprietary enterprise information.
Additional Amazon Q use instances introduced
Sivasubramanian introduced a preview of Amazon Q, the AWS pure language chatbot, in Amazon Redshift, which might present assist with writing SQL. Amazon Redshift with Amazon Q lets builders ask pure language questions, which the AI interprets right into a SQL question. Then, they will run that question and alter it as essential.
Plus, Amazon Q for information integration pipelines is now obtainable on the serverless computing platform AWS Glue for constructing information integration jobs in pure language.
Training and mannequin analysis instruments added to Amazon SageMaker
Sivasubramanian introduced the final availability of SageMaker HyperPod, a brand new distributed generative AI coaching functionality to cut back mannequin coaching time as much as 40%. SageMaker HyperPod can prepare generative AI fashions by itself for weeks or months, automating the duties of splitting information into chunks and loading that information onto particular person chips in a coaching cluster. SageMaker HyperPod consists of SageMaker’s distributed coaching pods, managed checkpoints for optimization, the power to detect and reroute round {hardware} failures. Other new SageMaker options embody SageMaker inference for quicker optimization and a brand new person expertise in SageMaker Studio.
Amazon SageMaker and Bedrock now have Model Evaluation, which lets prospects assess totally different basis fashions to seek out which is one of the best for his or her use case. Model Evaluation is offered in preview.
Vector capabilities and information administration instruments added to many AWS providers
Sivasubramanian introduced extra new instruments round vectors and information administration which can be appropriate for a wide range of enterprise use instances, together with generative AI.
- Vector Engine for OpenSearch Serverless is now usually obtainable.
- Vector capabilities are coming to Amazon DocumentDB and Amazon DynamoDB (out now in all areas the place Amazon DocumentDB is offered) and Amazon MemoryDB for Redis (now in preview).
- Amazon Neptune Analytics, an analytics database engine for Amazon Neptune or Amazon S3, is offered as we speak in sure areas.
- Amazon OpenSearch service zero-ETL integration with Amazon S3.
- AWS Clean Rooms ML, which lets organizations share machine studying fashions with companions with out sharing their underlying information.
“While gen AI still needs a strong foundation, we can also use this technology to address some of the big challenges in data management, like making data easier to use, making it more intuitive and making data more valuable,” Sivasubramanian mentioned.
Note: TechRepublic is masking AWS re:Invent nearly.
