AI2’s new open-source LLM may reset the definition of ‘open AI’

Long before ChatGPT, natural language AI researchers, many of them in academia, shared their research openly. The free-flowing exchange of information and innovation allowed the AI community at large to reproduce, validate, and criticize one another’s work. That all changed with the arrival of supersized LLMs like OpenAI’s GPT-4, when investors started pushing research labs to treat the details of their discoveries as valuable intellectual property—that is, to keep the underlying tech secret.

The Allen Institute for AI (AI2), the Seattle-based nonprofit started by Microsoft’s Paul Allen in 2014, wants to buck that trend. On Thursday, AI2 released a new large language model called OLMo 7B, and shared all the software components and training data that goes with it on GitHub and Hugging Face.

Hanna Hajishirzi [Photo: AI2]

“During this process we actually want to open up everything—the training data, the pretraining data, the source code, the details of the parameters, and so on,” says AI2 senior director of research Hanna Hajishirzi, who leads the OLMo project. “We are also releasing all these intermediate check points that we have obtained throughout training.”

The idea is to give the AI research community full visibility into a state-of-the-art large language model (LLM), which might enable it to advance natural language processing, and confront the problems with existing LLMs in a scientific way.

“We need to put in place a very clear methodology to evaluate how these models are working,” says AI2 COO Sophie Lebrecht, “and the only way to be able to do that is if we have full access to the data, so that we can go back and really understand how the model is behaving.”

In general, AI researchers are still struggling to attribute a specific output by an LLM to a particular piece of training data. Visibility into the reasoning of the model all the way from its training data through its decisions and outputs may help researchers make progress on that front. It could enable progress on other serious problems such as hallucinations and bias.

It’s also true that today’s LLMs are so big—and so expensive to train and operate—that many researchers are forced to use large closed models (via an API) from well-monied players like OpenAI or Google to conduct AI-assisted research. But in doing so they must take the output of those models as-is, with no way of understanding the “why” and “how” of the output.

“Being a researcher in the AI field and just working with APIs or closed models is like being an astronomer trying to research the Solar System and only having access to pictures of it from the newspaper,” Hajishirzi says.

Quoted in the company’s OLMo announcement is Meta chief AI scientist Yann LeCun, an outspoken proponent of open-sourcing new AI models. “The vibrant community that comes from open source is the fastest and most effective way to build the future of AI,” he said in the announcement, echoing a commonly used mantra.

Hajishirzi says Meta’s open-source Llama models have been extremely valuable, but even they aren’t completely open. “They have made the model open but still the data is not available, we don’t understand the connections starting from the data all the way to capabilities,” she says. “Also, the details of the training code is not available. A lot of things are still hidden.”

OLMo is considered a midsized model, with seven billion parameters (the synapse-like connection points in a neural network that contain weighted values). It was trained using two trillion tokens (words, word parts, or phrases).

Lebrecht points out that in an environment where AI researchers keep their discoveries secret, other researchers in academia or in other tech companies often end up trying to reengineer the work. The same work gets repeated, and that has major effects on the amount of power being used to run the servers, and on the carbon effects of that on the environment.

“By opening this up, these different research groups or different companies don’t need to do this siloed research,” Lebrecht says. “So when you open this up, we think it’s going to be huge in decarbonizing the impact of AI.”

https://www.fastcompany.com/91021305/ai2-new-open-source-llm?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Created 1y | Feb 1, 2024, 6:10:05 PM


Login to add comment

Other posts in this group

How AI is transforming corporate finance

The role of the CFO is evolving—and fast. In today’s volatile business environment, finance leaders are navigating everything from unpredictable tariffs to tightening regulations and rising geopol

Jul 5, 2025, 1:10:03 PM | Fast company - tech
Want to move data between Apple and Google Maps? Try this  workaround

In June, Google released its newest smartphone operating system, Android 16. The same month, Apple previewed its next smartphone oper

Jul 5, 2025, 10:40:07 AM | Fast company - tech
Tally lets you design great free surveys in 60 seconds

This article is republished with permission from Wonder Tools, a newsletter that helps you discover the most useful sites and apps. 

Jul 4, 2025, 1:50:03 PM | Fast company - tech
How China is leading the humanoid robots race

I’ve worked at the bleeding edge of robotics innovation in the United States for almost my entire professional life. Never before have I seen another country advance so quickly.

In

Jul 4, 2025, 9:20:03 AM | Fast company - tech
‘There is nothing that Aquaphor will not fix’: The internet is in love with this no-frills skin ointment

Aquaphor has become this summer’s hottest accessory.

The no-frills beauty staple—once relegated to the bottom of your bag, the glove box, or a bedside drawer—is now dangling from

Jul 3, 2025, 11:50:07 PM | Fast company - tech
Is Tesla screwed?

Elon Musk’s anger over the One Big Beautiful Bill Act was evident this week a

Jul 3, 2025, 5:10:05 PM | Fast company - tech