The IndexGuidesNews
AboutContribute

YuLan-Mini - A new highly open model

28 May 2025

At the European Open Source AI Index, we welcome any efforts to promote openness in the AI space. Most new models we add end up somewhere midway the index, often because they build on widespread models like Llama or Mistral that are not themselves very open to start with. Much to our delight, there is a genuinely new model in town that has overtaken good old BLOOMZ for the second-place spot.

YuLan-Mini is a new model by the Gaoling School of Artificial Intelligence. Claiming particularly good performance in math and code, the model is fully open according to nearly all of our openness measures. The model creators publish the data used to train their base model on HuggingFace, providing thorough information on the data mixture in an accompanying table. The model weights themselves are published under an open MIT license, and training procedures and code are documented both on GitHub and in a corresponding paper. The model itself is made available through Ollama for convenient use.

The detailed documentation of YuLan-Mini brings home the degree of scaffolding possible in the current open source AI landscape. For instance, to bootstrap maths abilities, YuLan-Mini uses Qwen 2.5 Math 7B Instruct as a teacher model; for instruction tuning, AllenAI's DOLMA dataset plays an important role; and for reward modeling the Skywork Reward model, built on public data, is used. YuLan-Mini exemplifies the continuing reliance of current models on large amounts of synthetic data, in a tradition that goes back to Alpaca's first GPT4-derived datasets. The prominence of synthetic data is something we follow with interest, as model makers have to walk a fine line between performance improvements and model collapse.

We commend the effort involved with open-sourcing this model to such a significant extent. Somewhat surprisingly, YuLan-Mini has received relatively little attention from the open-source community. With all versions of the model currently sitting at less than 150 downloads per month, we think there is a lot more potential for this model out there. We encourage anyone interested in open source generative AI to give this model a spin, peek under the hood, and learn from its high documentation standards.

Parameter descriptions:

Base Model Data
Are datasources for training the base model comprehensively documented and made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the end user interacts with comprehensively documented and made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the end user interacts with made freely available?
Training Code
Is the source code of dataset processing, model training and tuning comprehensively made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card available in standardized format that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Last updated 11 Jun 2025
OLMo by Ai2
OLMo-2-0325-32B
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
BLOOMZ by BigScience Workshop
BLOOM
Poro by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Whisper by OpenAI
Whisper-large-v3
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
Phi by Microsoft
Phi-4
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-32B-0414
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
Qwen by Alibaba Cloud
Qwen3-235B-A22B-Base
Granite by IBM
Granite-3.3-8B-Base
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Llama 4 by Meta
Llama 4 Maverick 17B 128E
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
BitNet by Microsoft
BitNet b1.58 2B4T
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-Small-9B-Base-R
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B

Supported by the Centre for Language Studies and the Dutch Research Council. Website design & development © 2024 by BSTN. This version of the index generated 11 Jun 2025, website content last updated 10 Jun 2025.