Amazon Bedrock now offers the AI21 Labs Jamba 1.5 models.
Jamba 1.5 family
We are pleased to inform that Amazon Bedrock now offers the potent new Jamba 1.5 family of large language models (LLMs) from AI21 Labs. With these models, long-context language capabilities have advanced significantly and are now more performant, fast, and useful for a variety of applications. Two variants in the Jamba 1.5 family are the Jamba 1.5 Mini and Jamba 1.5 Large. Both models allow the consumption of document objects, function calls, and a 256K token context window with structured JSON output.
Building foundation models and artificial intelligence (AI) systems for the enterprise is a specialty of AI21 Labs. By working strategically together, AI21 Labs and AWS are enabling clients in a variety of industries to develop, implement, and grow generative AI systems that address pressing issues and spur creativity. Customers may use LLMs in a safe setting to influence how people process information, communicate, and learn going forward by utilizing AI21 Labs’ cutting-edge, production-ready models in conjunction with Amazon’s specialized services and robust infrastructure.
What is Jamba 1.5?
The transformer model architecture and Structured State Space model (SSM) technology are combined in a novel hybrid architecture used by Jamba 1.5 models. With this novel method, Jamba 1.5 models can manage lengthy context windows up to 256K tokens while retaining the high-performance features of conventional transformer models. This hybrid SSM/transformer architecture is covered in greater detail in the whitepaper Jamba: A Hybrid Transformer-Mamba Language Model.
Amazon Bedrock now supports two new Jamba 1.5 models from AI21:
Jamba 1.5 Large: Jamba 1.5 For applications that demand high-quality results on both long and short inputs, Large is the perfect choice because it performs exceptionally well on complex reasoning tasks across all prompt lengths.
Jamba 1.5 Mini: Fast examination of large documents and data is made possible by Jamba 1.5 Mini’s low-latency processing of long prompts.
The Jamba 1.5 models’ main advantages are as follows:
- Extended context handling – Jamba 1.5 models, with their 256K token context length, can enhance the performance of enterprise applications, including extensive document summarizing and analysis, as well as agentic and RAG workflows.
- Multilingual: Hebrew, Arabic, German, Dutch, Spanish, French, Portuguese, Italian, and English are all supported.
- Friendly to developers: It can automatically handle structured JSON output, invoke functions, and process document objects.
- Efficiency and speed: AI21 evaluated the Jamba 1.5 models, reporting that the models outperform other models of similar sizes by up to 2.5X when it comes to inference on lengthy contexts.
Start using Jamba 1.5 models on Amazon Bedrock now
To begin using the new Jamba 1.5 models, navigate to the Amazon Bedrock console, select Model access from the pane on the bottom left, and submit a request to access Jamba 1.5 Small or Jamba 1.5 Large.
Select the Text or Chat playground from the left menu pane in the Amazon Bedrock dashboard to test the Jamba 1.5 models. Next, click Select model, choose AI21 as the category, and choose between Jamba 1.5 Mini and Jamba 1.5 Large for the model.
AWS SDKs can be used to access accessible models, and you can use a variety of programming languages to develop your apps.
For use cases like paired document analysis, compliance analysis, and lengthy document question-answering, the Jamba 1.5 models are ideal. They can effortlessly process lengthy or complex papers, evaluate data from several sources, and determine if passages adhere to particular rules. The AI21-on-AWS GitHub repository contains sample code. Visit AI21’s documentation to find out more about efficiently prompting Jamba models.
Presently accessible
Currently, the Jamba 1.5 model family from AI21 Labs is widely accessible in the US East (North Virginia) AWS Region’s Amazon Bedrock. Visit the AI21 Labs in Amazon Bedrock product and pricing pages to find out more.