Skip to content
AI21 Labs logo

AI21 Jamba 1.5 Mini

Playground
What are some common features of Gothic architecture?
What are some of the most famous works of Shakespeare?
Can you explain the concept of time dilation in physics?

Model navigation navigation

AI21 Labs

Jamba 1.5 Mini is a state-of-the-art, hybrid SSM-Transformer instruction following foundation model. It's a Mixture-of-Expert model with 52B total parameters and 12B active parameters. The Jamba family of models are the most powerful & efficient long-context models on the market, offering a 256K context window, the longest available.. For long context input, they deliver up to 2.5X faster inference than leading models of comparable sizes. Jamba supports function calling/tool use, structured output (JSON), and grounded generation with citation mode and documents API. Jamba officially supports English, French, Spanish, Portuguese, German, Arabic and Hebrew, but can also work in many other languages.

Model Developer Name: AI21 Labs

Model Architecture

Jamba 1.5 Mini is a state-of-the-art, hybrid SSM-Transformer instruction following foundation model

Model Variations

52B total parameters and 12B active parameters

Model Input

Model inputs text only.

Model Output

Model generates text only.

Model Dates

Jamba 1.5 Mini was trained in Q3 2024 with data covering through early March 2024.

​​​Model Information Table

Name Params Content Length
Jamba 1.5 Mini 52B (12B active) 256K
Jamba 1.5 Large 398B (94B active) 256K

About

A 52B parameters (12B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.
Context
262k input · 4k output
Training date
Undisclosed
Rate limit tier
Provider support

Languages

 (7)
English, French, Spanish, Portuguese, German, Arabic, and Hebrew