Getty Images

Ai2's new model more open than DeepSeek

The AI research lab released Tülu 3 in November. The 405B version was created with reinforcement learning from variable rewards and was technically challenging to create.

The Allen Institute for Artificial Intelligence released the 405B parameter of its latest language model this week, claiming Tülu 3 405B, performs better than Chinese startup DeepSeek V3 and OpenAI GPT4o.

The open source model was created using reinforcement learning from variable rewards (RLVR). This approach involves training the model to enhance specific skills, such as mathematical problem-solving and instruction following.

The AI research lab first introduced Tülu 3 in November, using the same RLVR approach and method. Despite its success in scaling the method with Tülu 3 405B, Ai2 had some technical challenges with the model. For one, Tülu 3 405B required 256 GPUs running in parallel; hyperparameter tuning was limited as a result of those computational costs, , the research lab said.

Ai2's release of Tülu 405B comes at a time when Chinese startup DeepSeek has disrupted both the U.S. and Chinese AI markets with its reasoning model DeepSeek R. Its DeepSeek V3 model was released last year.

Tülu 3 405B was also released on the same day that French startup Mistral AI released open source model Mistral Small 3, a 24-parameter model under the Apache 2.0 license.

Innovation and openness

DeepSeek, Mistal Small 3 and Tülu 3 405B all show the continual growth of the open source market and the continued innovation in the AI market.

"We're seeing the iterative and evolutionary change ... morphing of these models," said Mark Beccue, an analyst at Enterprise Strategy Group, now part of Omdia.

While it's important to see the models get better in terms of performance and accuracy, Ai2's strength comes from its openness, Beccue said.

They were very open about this wasn't cheap.
Mark BeccueAnalyst, Enterprise Strategy Group, now part of Omdia

"They were very open about this wasn't cheap," he said.

This is different than DeepSeek's reasoning model R1, which the Chinese startup claims is open source.

However, many experts are questioning whether R1 is truly open source because the data it was trained on and the components used to build it are not publicly available. There are also questions about the validity of DeepSeek R1 being cost efficient.

In contrast with DeepSeek and others, Ai2 is known for releasing not just its training code and models but also its datasets.

"Ai2's fully open approach ... ensures users can easily customize their pipeline for everything from data selection through evaluation," said Constellation Research analyst Andy Thurai.

An open approach is also better for accuracy, Beccue said.

"I hope Ai2, this kind of model really takes off," he said.

Gaining traction

However, Ai2 hasn't had much traction in the enterprise world, Thurai said.

"Their messaging, marketing and propaganda are very low-key," he said. "Their models are equally performing, but their model news is literally two blogs, versus the DeepSeek model release was a newsworthy item all the way up to the moon."

The lack of traction in the enterprise world may be due to the motivation of the research lab, Beccue said. As a research and nonprofit, Ai2’s motivation may not necessarily be to use what it creates to advance its own profit, compared with other creators of LLMs, Beccue said.

"There's kind of some purity to that," he said.

The new Tülu 3 405B model is available to test on Ai2's chatbot web app. Its code is also available on GitHub and Hugging Face and it's hosted on Google Cloud.

Esther Shittu is an Informa TechTarget news writer and podcast host covering artificial intelligence software and systems.

Dig Deeper on AI technologies