’ “ ℎ ”

With Pixtral 12B, we’re looking at a smaller parameter count compared to giants like – , . The real test will be how ’ , , .

’ – ?

As far as I can tell, Mistral AI stance is that this is their latest multimodal large language model (LLM). At 12 billion parameters, Pixtral might seem small compared to models like GPT-4. But what’s exciting is how these smaller models continue to surprise us with their efficiency and speed.

The introduction of Pixtral 12B also comes with some notable updates. ‘ : . , . _ , and . _ . These tokens enhance the model’s image processing capabilities, making it easier to embed and analyze visual data alongside text prompts, a interesting step forward in AI’s ability to handle multimodal tasks.

Pics from a Reddit thread. I’ll add info in the comments.

#ai #genai #multimodal #news

: The views within any of my posts, or newsletters are not those of my employer or the employers of any contributing experts. this? Feel free to reshare, repost, and join the conversation.

Picture of Doug Shannon

Doug Shannon

Doug Shannon, a top 50 global leader in intelligent automation, shares regular insights from his 20+ years of experience in digital transformation, AI, and self-healing automation solutions for enterprise success.