European Open Source AI Index
DatabaseNewsGuidesAboutContribute
Guide
by Mark Dingemanse
12 October 2024

Generative AI models have many moving parts. This guide provides a survey of the most important openness dimensions by discussing two models both self-billed as "open source". BloomZ was introduced by the BigScience Workshop team in May 2023 as an early open source large language model; Llama 3.1 (Llama for short) was introduced by Facebook Research as “the next generation of our open source large language model”. However, a glance at the openness scores below shows that the models differ quite a lot in terms of overall openness.

Last updated 09 April 2026
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
BLOOMZ by BigScience Workshop
BLOOM
OLMo by Ai2
Olmo-3-1125-32B
Apertus by Swiss AI Initiative
Apertus-70B-2509
SmolLM by HuggingFace
SmolLM3-3B-Base
mT0 by BigScience Workshop
mT5-XXL
Amber by LLM360
Amber
Pythia by EleutherAI and Together Computer
Pythia-6.9B
Open Assistant by Open Assistant
Pythia-12B
Lucie by OpenLLM-France
Lucie-7B
Instella by AMD
Instella-3B
EuroLLM by UTTER
EuroLLM-22B-2512
K2 by LLM360
K2-V2-Instruct
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Skywork-OR1 by Skywork
DeepSeek-R1-Distill-Qwen-32B
Omnilingual ASR by Meta
Omnilingual ASR
MobileLLM by Meta
MobileLLM-R1-950M-base
BTLM by Cerebras
BTLM-3B-8K-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
Whisper by OpenAI
Whisper-large-v3
Teuken by OpenGPT-X
Teuken-7B-base
T5 by Google AI
T5
Eurus by OpenBMB
Mixtral-8x22B-v0.1
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Poro by AMD Silo AI and TurkuNLP and High Performance Language Technologies (HPLT)
Llama-Poro-2-70B-Base
OpenChat by OpenChat
Meta-Llama-3-8B
Neo by Multimodal Art Projection
Neo-7B
Guru by LLM360
Guru-32B
BERT by Google AI
unspecified
Dolly by Databricks
Pythia-12B
Vicuna by LMSYS
Vicuna-13B
Tülu by Ai2
Llama-3.1-405B
TildeOpen by Tilde.ai
TildeOpen-30b
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
OpenMoE by Zheng Zian
OpenMoE-8B
Occiglot by Occiglot
Occiglot-7B-EU5
Llama Nemotron by NVIDIA
Llama-3.3-70B-Instruct
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
Baguettotron by PleIAs
Baguettotron
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
WizardLM by Microsoft and Peking University
LLaMA-7B
SynLogic by Minimax AI
SynLogic-32B
Phi by Microsoft
Phi-4
OpenELM by Apple
OpenELM-3B
NeuralChat by Intel
Mistral-7B-v0.1
DeepHermes by Nous Research
Llama-3.1-70B
Pharia by Aleph Alpha Research
Pharia-1-LLM-7B
minChatGPT by Ethan Yanjia Li
GPT2-Medium
DeepSeek V3.2 by DeepSeek
DeepSeek-V3.1-Base
Yi by 01.AI
Yi-34B
XBai-04 by Yuan Shi Technology
Qwen3-32B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Saul by Equall
Mixtral-8x22B-v0.1
Hunyuan by Tencent
Hunyuan-7B-Pretrain
Apriel by ServiceNow
Apriel-1.5-15b-Thinker
MiMo by Xiaomi
MiMo-V2-Flash-Base
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Xwin-LM by Xwin-LM
Llama-2-13B
Geitje by Bram Vanroy
Mistral-7B-v0.1
GPT OSS by OpenAI
unspecified
Claire by OpenLLM-France
Falcon-7B
BELLE by KE Technologies
Llama-2-13B
UltraLM by OpenBMB
Llama-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Solar by Upstage AI
unspecified
Marco by Alibaba
Marco-LLM-GLO
MPT by Databricks
MPT-30B
Intern-S1 by Shanghai AI Laboratory
Intern-S1-Pro
Granite by IBM
Granite-4.0-H-Small-Base
Bielik by SpeakLeash AI
Bielik-11B-v3-Base-20250730
Viking by Silo AI and TurkuNLP and High Performance Language Technologies (HPLT)
unspecified
Mistral NeMo by Mistral AI and NVIDIA
Mistral-NeMo-12B-Base
Starling by NexusFlow
Llama-2-13B
Seed-OSS by ByteDance
Seed-OSS-36B-Base
Nanbeige by Nanbeige LLM lab
Nanbeige4-3B-Base
LongAlign by Zhipu AI
Llama-2-13B
Ling by Inclusion AI
Ling-2.5-1T
Kimi K2.5 by Moonshot AI
unspecified
Gemma by Google AI
Gemma-3-27B-PT
Falcon by Technology Innovation Institute
Falcon-H1-7B-Base
Llama 4 by Meta
Llama-4-Maverick-17B-128E
dots.llm1 by RedNote
dots.llm1.base
Stanford Alpaca by Stanford University CRFM
Llama-7B
Jamba by AI21
AI21-Jamba2-Mini
GLM by Zhipu AI
GLM-5
Llama 3.1 by Meta
Llama-3.1-405B
XVERSE by Shenzhen Yuanxiang Technology
XVERSE-MoE-A4.2B
TeleChat by Tele-AI
unknown
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
RWKV by BlinkDL/RWKV
unspecified
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Mistral Large 3 by Mistral AI
Mistral-Large-3-675B-Base-2512
LFM2 by Liquid AI
unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Base
FastChat-T5 by LMSYS
Flan-T5-XL
EXAONE by LG
unspecified
Crystal by LLM360
Crystal
Cogito by DeepCogito
DeepSeek-V3-Base
BitNet by Microsoft
unspecified
StableVicuna by CarperAI
LLaMA-13B
Llama 2 by Meta
Llama-2-70B
LeoLM by LAION
Llama-2-70B
Koala by BAIR
Llama-13B
XGen by Salesforce
XGen-Small-9B-Base-R
Reka Flash by Reka AI
unknown
Jais by G42
unspecified
Llama 3.3 by Meta
Llama-3.3-70B
Stable Beluga by StabilityAI
Llama-2-70B
Minimax-M2.5 by Minimax AI
MiniMax-Text-01?
Llama-Sherkala by G42
Llama-3.1-8B
Llama 3 by Meta
Meta-Llama-3-70B
Qwen by Alibaba
unspecified
Gemma Japanese by Google AI
Gemma-2-2B
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
Command A by Cohere AI
Command A?
Grok 2 by xAI
unknown
Lumo AI by Proton
Undisclosed

Hovering over the openness indicators below makes visible the 14 dimensions by which we judge model openness. In this guide, we summarise the dimensions in terms of three areas: availability, documentation and access.

Availability

When it comes to open code, we find that BloomZ makes available source code for training, fine-tuning and running the model, while for Llama none of the model's source code is made available, only scripts for running the model are shared. The LLM data underlying the base model is documented in great detail by BloomZ, while for Llama only the vaguest details are provided in a corporate preprint: “a new mix of data from publicly available sources, which does not include data from Meta's products or services”. The statement is clearly designed to minimise legal exposure.

Both systems make available model weights, though for Llama access is restricted through a consent form. The training data for instruction tuning (RL data) is described and documented by BloomZ as consisting of xP3 (Crosslingual Public Pool of Prompts); for Llama, the corporate preprint notes that fine-tuning was done based on “a large dataset of over 1 million binary comparisons based on humans applying our specified guidelines, which we refer to as Meta reward modeling data”, and which remains undisclosed. (The same preprint mentions that for evaluation, Meta did build on several RLHF datasets openly shared by others.) Model weights for the instruction-tuned version (RL weights) are made openly available by BloomZ, while for Llama they require an access request.

Documentation

The BloomZ code is not just available, it is also well-documented. For Llama on the other hand, no no documentation of source code is available (as the source code itself is not open). Model architecture is described for BloomZ in multiple scientific papers and supported by a github repository of code and recipes on HuggingFace; for Llama, the architecture is described in less detail and scattered across corporate websites and a preprint.

BloomZ's multiple preprints document data curation and fine-tuning in great detail; in contrast, Llama's single preprint offers fewer details and appears strategically vague on crucial details (for instance, training datasets and instruction tuning). The scientific documentation of BloomZ also includes multiple peer-reviewed papers, including one of the very few scientifically vetted sources of data on the energy footprint of training large language models. No peer-reviewed papers providing scientific documentation or evaluation of Llama are known currently.

The two systems also differ in terms of the sytematicity of documentation, as measured by the availability of model cards and data sheets: industry-standard models for providing metadata on architecture, training, and evaluation. For Bloom, these system cards provide basic details alongside extensive cross-references to other documentation on training data, training approach, model architecture, fine-tuning and responsible use. In contrast, the Llama model card only provides minimum detail and none whatsoever on training data. A data sheet is only available for BloomZ. This means that for Llama, there is no documentation of training datasets whatsoever — a prime example of a strategy described by Birhane et al. as a tactical template of “(non)declaring the training dataset information”.

Access

Access presents a mixed picture for both of the models. Both are primarily intended for local deployment, and are additionally available through various application programming interfaces (APIs).

Neither model is fully released under an OSI-approved open source license, but the licensing details are interestingly different. The BloomZ source code is released under Apache 2.0, a maximally open license with a minimum of restrictions. The model weights on the other hand are released under the Responsible AI License (RAIL). Llama releases no source code, as we saw above; meanwhile the model weights are released under a bespoke "Meta Community License".

Both licences aim to restrict harmful use cases, but there is a key difference in the constraints they put on representing model outputs. RAIL stipulates that a user may not “generate content without expressly and intelligibly disclaiming that the text is machine-generated”. the Meta Community License for Llama on the other hand stipulates that a user may not “represent that Llama 2 outputs are human-generated”. This is a much lower bar, because it leaves open a wide swathe of use cases where there may not be the explicit claim of human-generated output, but merely a strong implication.

In closing

This brief guide offers a walkthrough of the main dimensions of openness. As we see, the open source claim of BloomZ is well-founded. Llama on the other hand really cannot be called "open source" under any reasonable definition of the term. It is at best open weights, and is closed in almost all other aspects. Llama, in all currently available versions, is a prime example of a model that claims openness benefits by merely providing access to its most inscrutable element: model weights.

The models can also be compared side by side.

This guide incorporates some text and data from the following paper:

  • Liesenfeld, Andreas, and Mark Dingemanse. 2024. ‘Rethinking Open Source Generative AI: Open-Washing and the EU AI Act’. In The 2024 ACM Conference on Fairness, Accountability, and Transparency (FAccT ’24). Rio de Janeiro, Brazil: ACM. doi: 10.1145/3630106.3659005
Supported by the Centre for Language Studies and the Dutch Research Council. Website design & development © 2024 by BSTN. This version of the index generated 09 April 2026, website content last updated 11 March 2026.