In recent years, the field of AI has grown dramatically. Besides the introduction of ChatGPT and the subsequent revolution in text-based LLMs, there has been a revolution of no lesser significance in image-based diffusion models. The surge of new models can make it challenging to see what types of model are available and how current models should be categorized. Here we provide an overview of the most common types of AI models currently available, with the aim of helping the reader to get a better grasp on the AI landscape as a whole.
Throughout this blog post, we provide references to example models in our index, if available, using a grid view.
Are datasources for training the base model comprehensively documented and freely made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the enduser interacts with comprehensively documented and freely made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the enduser interacts with made freely available?
Training Code
Is the source code of datasource processing, model training and tuining comprehensively and freely made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card in standardized format available that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation are available?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Poro-34B by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Phi by Microsoft
Phi-4
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-9B
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
QwQ-32B by Alibaba Cloud
Qwen2.5-32B
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-7B-4K-Base
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
Granite by IBM
Granite-3.1-8B-Base
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B
Let's start with the models which initiated the recent revolution in the field of AI; text-based LLMs. It is useful to think of text-based models such as ChatGPT as constructed in (at least) two phases. First, a base LLM is developed which stores a model of human language and uses it to generate text. Subsequently, this model is 'fine-tuned' (slightly modified) in various ways to serve as an effective chatbot.
A base LLM (sometimes described as a foundation LLM) is a text prediction engine that is trained to generate probabilities of possible next tokens given an input prompt. Recent text-based LLMs are typically trained on vast amounts of data, including countless websites, books and other text sources. By statistically representing the relationships between billions of tokens, they can generate plausible continuations of sequences of tokens. Base LLMs are often trained on supercomputers, since the amount of processing power required is vast.
Though a base LLM is excellent for text prediction tasks, it does not straightforwardly allow for carrying on conversations. For this, a derivative model is necessary in the form of an instruct model. Instruct models are created by taking an existing base model, and repeatedly showing it examples from human text-based interactions. This allows the model to generate patterns mirrorring some aspects of the source interactions, opening up the possibility for people to treat it as a conversational participant.
Chat models are instruct models which have received additional fine-tuning to avoid harmful conversation topics and to serve as helpful assistants. By further tuning a model on examples of helpful AI conversations and showing examples where the model rejects discussing harmful topics, chat models can be made to pick up on the pattern that they should output text in a helpful tone and reject harmful topics.
It is worth noting that although the terms 'chat model' and 'instruct model' are usually used in the sense described above, the terms are also occasionally used interchangeably. A model may be advertised as an instruct model when it has in fact undergone safety fine-tuning, and a model which has only undergone instruction tuning may sometimes be described as a chat model. The ambiguity with regards to this terminology is partially due to the inherent complexity of the underlying subject matter, and partially due to frequent attempts to combine the two tuning stages into one. In general, it is best to inspect the specifics of a chat- or instruct-model on a case-by-case basis.
Fine-tuning a base model into an instruct model and subsequently a chat model requires far less computing power than training the foundation model from scratch. Typically, only a few million instructions are needed and the amount of processing power required is only slightly more than the amount required for running the model in the first place. As such, the barrier of entry for deriving chat models from base models is far lower than the barrier of entry for constructing base models in the first place. This means that many more players can participate in the construction of chat- and instruct-models than in the construction of base models.
In our index, we include chat- and instruct-models under the 'text' mode label. For each model, we also list the underlying base models.
Are datasources for training the base model comprehensively documented and freely made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the enduser interacts with comprehensively documented and freely made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the enduser interacts with made freely available?
Training Code
Is the source code of datasource processing, model training and tuining comprehensively and freely made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card in standardized format available that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation are available?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Poro-34B by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Phi by Microsoft
Phi-4
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-9B
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
QwQ-32B by Alibaba Cloud
Qwen2.5-32B
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-7B-4K-Base
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
Granite by IBM
Granite-3.1-8B-Base
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B
Reasoning models are the most recent types of models which have emerged in the AI space, exemplified by models such as DeepSeek-R1 and basing themselves off techniques pioneered by models such as Eurus. Such reasoning models are better able to handle multi-step reasoning tasks by breaking them up into separate steps.
It is worth noting that reasoning LLMs do not strictly 'reason' in the common sense. Instead, a bit like chat models, they are tuned to do particularly well at a task known as chain-of-thought reasoning, in which they reproduce patterns of stepwise reasoning that they have been exposed to in training data. Since the output of intermediate steps becomes part of the context window, this can improve the final answer. The stepwise procedure gives the appearance of the LLM thinking and reflecting on a prompt.
Are datasources for training the base model comprehensively documented and freely made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the enduser interacts with comprehensively documented and freely made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the enduser interacts with made freely available?
Training Code
Is the source code of datasource processing, model training and tuining comprehensively and freely made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card in standardized format available that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation are available?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Poro-34B by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Phi by Microsoft
Phi-4
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-9B
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
QwQ-32B by Alibaba Cloud
Qwen2.5-32B
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-7B-4K-Base
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
Granite by IBM
Granite-3.1-8B-Base
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B
Code LLMs are LLMs specializing in handling code. There are two main ways of creating a code LLM. Firstly, it is possible to further train a regular base, chat or instruct model using code data (possibly adding some code-integrating instructions) to arrive at a model which can understand and generate code in a conversational setting. The model this produces is akin to the 'chat' functionality of GitHub Copilot, with the model serving as a helpful assistant for aiding in code-related tasks.
The second way of creating a code LLM is to train a model from scratch using code data. This results in a model which does not necessarily understand human language, but understands how to complete code very well. As such, models produced in this way are primarily useful as code autocompletion agents.
Code LLMs are usually tuned on multiple programming languages, allowing for a great deal of versatility. However, it is not uncommon to see further fine-tuning of a code model to a specific language (e.g. Python), to allow it to perform particularly well when programming in one specific setting.
Math LLMs are large language models fine-tuned to perform well on mathematical reasoning tasks. Similarly to code LLMs, a few approaches are possible.
Firstly, it is possible to tune a pre-existing large language model to perform particularly well at mathematical reasoning. Existing approaches have tuned base, chat-tuned, instruct-tuned, and even code-tuned LLMs to provide good mathematical reasoning. However, all of these approaches generally converge to a model that specializes in solving and reasoning about mathematical problems in natural language.
Next, it is possible to tune a language model in the same way one would tune a code completion LLM, making it specialize in completing mathematical proofs using a proof assistant. This allows for more logically sound reasoning; however, it also requires translating a problem to strictly formal natural language.
Lastly, recent approaches have looked into an LLM which somewhat combines the former two approaches; being able to both understand and think about mathematical problems in natural language and translate the corresponding problem to the language of proof assistants. On the whole, the field of math LLMs remains quite experimental.
Agentic LLMs are LLMs specifically tuned to interface with tools or to perform web-based tasks. Given that both the approaches taken and the applications for which agentic LLMs are designed vary widely, it is difficult to say anything about this group in a general sense. However, a few common facts hold. Firstly, agentic LLMs tend to specialize in either augmenting an LLM to interface well with external tools (e.g. a way of constructing and running programs), or in acting as sophisticated web-based agents. Agentic LLMs also tend to be quite experimental, with each adopting novel techniques for optimizing agentic actions.
Given the emergent nature of agentic models, there are many potential barriers that still need to be addressed. Most notably, agentic LLMs are usually not sufficiently open-source and potentially pose inherent security risks. Future research remains to demonstrate the potential utility of such models and investigate how they can be constructed in an open manner.
Although text-based LLMs represent the most popular use case and have seen the most activity in terms of model development, there is also a wide variety of AI models capable of generating audiovisual media of various kinds. Here we highlight the most notable types.
Are datasources for training the base model comprehensively documented and freely made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the enduser interacts with comprehensively documented and freely made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the enduser interacts with made freely available?
Training Code
Is the source code of datasource processing, model training and tuining comprehensively and freely made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card in standardized format available that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation are available?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Poro-34B by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Phi by Microsoft
Phi-4
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-9B
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
QwQ-32B by Alibaba Cloud
Qwen2.5-32B
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-7B-4K-Base
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
Granite by IBM
Granite-3.1-8B-Base
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B
Image diffusion models are the powerhouses which fueled the recent revolution in AI image generation. Notable members include Midjourney and Stable Diffusion. Image diffusion models operate by attempting to learn how to extract images in the target context from random noise. An image diffusion starts with a completely noisy image and, by subsequently predicting which noise was added, attempts to reconstruct a fully non-noisy image. Consumer-oriented image diffusion models typically condition their output on text, allowing for generating an image a particular way based on a text prompt. The way image diffusion models are trained is typically to feed them images along with text data describing the image, allowing the model to both learn how to generate the images and how to connect the images to the text.
The architecture of an image diffusion model is amenable to many modifications, allowing for, for instance, generating an image based on a different image (image modification), or for generating an image based off a sketch or texture map (ControlNet). Image models are often fine-tuned to output images within specific genres, such as cartoons or paintings.
Are datasources for training the base model comprehensively documented and freely made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the enduser interacts with comprehensively documented and freely made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the enduser interacts with made freely available?
Training Code
Is the source code of datasource processing, model training and tuining comprehensively and freely made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card in standardized format available that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation are available?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Poro-34B by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Phi by Microsoft
Phi-4
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-9B
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
QwQ-32B by Alibaba Cloud
Qwen2.5-32B
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-7B-4K-Base
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
Granite by IBM
Granite-3.1-8B-Base
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B
Video-generating models are often variations on the theme of image-generating models. The key challenge for video models compared to image-generating models is to provide consistency between frames, which they accomplish by conditioning subsequent frames on previous frames in various ways. Currently, video-generating models run into issues when attempting to generate longer videos, videos of variable length, and videos of varying resolution. It is also an open question how well a video-generating model can be constructed using solely open-source data. In general, though video-generating models have already seen some adoption, for now they remain quite experimental.
An aspirational fine-tune of the video model is the world model. World models seek to ground the video they generate into a virtual world and allow for some degree of control, creating a video-game-like experience. The aim of world models is to allow for the generation of grounded digital worlds. Though some large players in the AI space have already committed to the development of world models, in general this type of model still remains very experimental.
Are datasources for training the base model comprehensively documented and freely made available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Data
Are datasources for training the model that the enduser interacts with comprehensively documented and freely made available?
Base Model Weights
Are the weights of the base models made freely available? In case a distinction between base (foundation) and end (user) model is not applicable, this mirrors the end model data entries.
End User Model Weights
Are the weights of the model that the enduser interacts with made freely available?
Training Code
Is the source code of datasource processing, model training and tuining comprehensively and freely made available?
Code Documentation
Is the source code of datasource processing, model training and tuning comprehensively documented?
Hardware Architecture
Is the hardware architecture used for datasource processing and model training comprehensively documented?
Preprint
Are archived preprint(s) are available that detail all major parts of the system including datasource processing, model training and tuning steps?
Paper
Are peer-reviewed scientific publications available that detail all major parts of the system including datasource processing, model training and tuning steps?
Modelcard
Is a model card in standardized format available that provides comprehensive insight on model architecture, training, fine-tuning, and evaluation are available?
Datasheet
Is a datasheet as defined in "Datasheets for Datasets" (Gebru et al. 2021) available?
Package
Is a packaged release of the model available on a software repository (e.g. a Python Package Index, Homebrew)?
API and Meta Prompts
Is an API available that provides unrestricted access to the model (other than security and CDN restrictions)? If applicable, this entry also collects information on the use and availability of meta prompts.
Licenses
Is the project fully covered by Open Source Initiative (OSI)-approved licenses, including all data sources and training pipeline code?
Poro-34B by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Poro-34B
mT0 by bigscience-workshop
mT5-XXL
Pythia by EleutherAI, Together Computer
Pythia-6.9B
Open Assistant by LAION-AI
Pythia-12B
Amber by LLM360
Amber
YuLan by Gaoling School of Artificial Intelligence
YuLan-Mini
K2 by LLM360
K2
SmolLM by HuggingFace
SmolLM2-1.7B
OpenChat by OpenChat
Meta-Llama-3-8B
Arabic StableLM by StabilityAI
StableLM-2-1.6B
Intestella by AMD
Instella-3B
Dolly by Databricks
Pythia-12B
Tülu by Ai2
Llama-3.1-405B
T5 by Google AI
T5
RedPajama by Together Computer
RedPajama-INCITE-7B-Base
Phi by Microsoft
Phi-4
Neo by Multimodal Art Projection
Neo-7B
BERT by Google AI
BERT
AquilaChat by Beijing Academy of Artificial Intelligence
Aquila2-70B-Expr
DeepSeek V3 by DeepSeek
DeepSeek-V3-Base
Yi by 01.AI
Yi-34B
Teuken by openGPT-X
Teuken-7B-base
Salamandra by Barcelona Supercomputing Center
Salamandra-7B
NeuralChat by Intel
Mistral-7B-v0.1
MPT by Databricks
MPT-30B
Lucie by OpenLLM-France
Lucie-7B
GPT-SW3 by AI Sweden
GPT-SW3-6.7B-V2
GPT-NeoXT by Together Computer
GPT-NeoX-20B
Fietje by Bram Vanroy
Phi-2
BTLM by Cerebras
BTLM-3B-8K-Base
Pharia by Aleph Alpha Research
Pharia 1 LLM 7B
minChatGPT by Ethan Yanjia Li
GPT2
Eurus by OpenBMB
Mixtral-8x22B-v0.1
Xwin-LM by Xwin-LM
Llama-2-13B
Vicuna by LMSYS
LLaMA
OpenELM by Apple
OpenELM-3B
Occiglot by Occiglot
Occiglot-7B-EU5
Mistral by Mistral AI
Mistral-Large-2411
GLM by Zhipu AI
GLM-4-9B
Falcon by Technology Innovation Institute
Falcon3-10B-Base
Minerva by Sapienza Natural Language Processing Group
Minerva-7B-base-v1.0
DeepSeek R1 by DeepSeek
DeepSeek-V3-Base
Zephyr by HuggingFace
Mixtral-8x22B-v0.1
QwQ-32B by Alibaba Cloud
Qwen2.5-32B
InternLM by Shanghai AI Laboratory
InternLM3-8B
CT-LLM by Multimodal Art Projection
CT-LLM-Base
Mistral NeMo by Mistral AI, NVIDIA
Mistral NeMo
WizardLM by Microsoft & Peking University
LLaMA-7B
Starling by NexusFlow
Llama-2-13B
Saul by Equall
Mixtral-8x22B-v0.1
BELLE by KE Technologies
Llama-2-13B
Airoboros by Jon Durbin
Qwen1.5-110B
Gemma by Google AI
Gemma-3-27B-PT
Geitje by Bram Vanroy
Mistral 7B
Marco by Alibaba
Marco-LLM-GLO
Viking by Silo AI, TurkuNLP, High Performance Language Technologies (HPLT)
Viking-33B
UltraLM by OpenBMB
LLaMA2
Llama 3.1 by Meta
Meta Llama 3
OpenMoE by Zheng Zian
OpenMoE-8B
Command-R by Cohere AI
C4AI-Command-R-V01
Stanford Alpaca by Stanford University CRFM
Llama-7B
StripedHyena by Together Computer
StripedHyena-Hessian-7B
Stable Beluga by Stability AI
LLaMA2
LongAlign by Zhipu AI
Llama-2-13B
Claire by OpenLLM-France
Falcon-7B
Llama 3.3 by Meta
Llama 3.3 70B
Koala by BAIR
unspecified
RWKV by BlinkDL/RWKV
RWKV-x070-Pile-1.47B-ctx4096
Persimmon by Adept AI Labs
Persimmon-8B-Base
OPT by Meta
OPT-30B
Nanbeige by Nanbeige LLM lab
Unknown
Infinity-Instruct by Beijing Academy of Artificial Intelligence
Llama-3.1-70B
H2O-Danube by H2O.ai
H2O-Danube3.1-4B-Chat
FastChat-T5 by LMSYS
Flan-T5-XL
Crystal by LLM360
Crystal
Baichuan by Baichuan Intelligent Technology
Baichuan2-13B-Base
StableVicuna by CarperAI
LLaMA
Llama 3 Instruct by Meta
Meta Llama 3
XGen by Salesforce
XGen-7B-4K-Base
Solar by Upstage AI
LLaMA2
Llama-Sherkala by G42
Llama-3.1-8B
Jais by G42
Llama-2-70B
Hunyuan by Tencent
Hunyuan-A52B-Pretrain
Granite by IBM
Granite-3.1-8B-Base
DeepHermes by Nous Research
Llama-3.1-8B
LLaMA2 Chat by Meta
LLaMA2
Snowflake Arctic by Snowflake
Snowflake-Arctic-Base
Minimax-Text by Minimax AI
MiniMax-Text-01
Gemma Japanese by Google AI
Gemma-2-2B
Audio models are models for generating music and other auditory information. By and large they are equally as experimental as video models, though they have arguably seen slightly more adoption. Audio models are by and large still quite closed, leading to difficulty establishing their typical construction. Nonetheless, they have been shown capable of generating high-quality melodies, vocals, instrument tracks, and even entire songs.
Multimodal LLMs are perhaps the most varied of all LLM categories. Very broadly, multimodal LLMs seek to incorporate information from modalities other than text in order to enhance the utility of LLMs. The simplest and most common variant of a multimodal LLM is an image-text-to-text model, which can process images as well as text. However, many variants of multimodal LLMs exist. Notably, some multimodal models have grown to allow for both the processing and generation images and text, and one has proven capable of processing text, images, video, and audio and generating both text and audio.
The exact recipe for constructing a multimodal LLM remains largely in flux, with models processing and generating their target data in many different ways. A common approach, however, is to take a pre-existing LLM and further train it to be able to handle information across a range of target modalities.
Though this overview seeks to have outlined the most common types of AI models, many more exotic and narrow types still exist. The AI space is vast, and the process of navigating it can feel more like exploration than the plotting out of any particular route. We hope this overview of common model types provides readers with a useful view of the overall landscape, just as our index helps them to navigate questions of openness and transparency.