The European Open-Source AI Index collects information on model openness, licensing, and EU regulation of generative AI systems and providers.
The index is a non-profit public resource hosted at the Centre of Language and Speech Technology, Radboud University, The Netherlands, maintained by a small team of academics and community members.
Andreas Liesenfeld is co-founder of the index and Assistant Professor at Radboud University where he researches the societal and regulatory implications of emerging language technologies. His current work focuses on technology assessment, open source generative AI, and comparative AI regulation. At the Centre for Language Studies (CLS), Andreas Liesenfeld is a member of the language technology expertise group and, in collaboration with the Netherlands eScience center, develops open source research software for natural language processing.
Mark Dingemanse is co-founder of the EU OSAI index and Associate Professor at Radboud University. His current research looks at the intersection of language and technology, from the artisanal tools that people use in everyday conversation to the artificial techniques of large language models, voice user interfaces, and "AI". He directs the research programme Futures of Language funded by a VICI talent grant from the Dutch Research Council (2024-2029).
Dick Blankvoort is a student at Radboud University and a researcher at the Open Source AI Index. His primary interest lies in the field of Data Science, focusing both on the applications of generative AI and on their mathematical underpinnings. He is currently pursuing his internship at OSAI along with following an honours lab about applications of generative AI in education, as well as teaching a course on Computer Algebra.
Nityaa Kalra has no bio yet
Adem Kaya
Ada Lopez
Open-source software and systems are a crucial building block of innovative technology ecosystems and provide an important alternative to proprietary, closed systems for users and builders of generative AI systems. True openness fosters innovation that not only serves commercial interests, but also contributes to knowledge access, free education, and a more informed public debate about this emerging technology.
In 2025, the AI landscape will be shaken up by the enforcement of the EU AI Act, the world's first comprehensive AI law, with a projected impact comparable to GDPR. This regulation pays special attention to open-source systems and regulates ‘open’ systems differently. But the EU AI Act neither defines openness nor specifies how open-source licenses relate to openness. This work will be carried out by the future EU AI Office.
The unclear definition of openness in the domain of generative AI creates loopholes for overclaiming openness (by ie. using an open source license for a single aspect) and the risk of unintended claims to exemptions based on open-source status under EU law. A practice known as ‘open-washing’.
We created an evidence-based framework that distinguishes 14 dimensions of openness, from training datasets to scientific and technical documentation and from licensing to access methods. We published our research in a paper. The data is continuously updated and publicly available via our GitHub repository. Contributions are encouraged. The website you’re currently reading is providing a user-friendly interface for the data from the Github repository and gives extra contextual information to a wider audience.
Evidence-based openness assessment can help foster a generative AI landscape in which models can be effectively regulated, model providers can be held accountable, scientists can scrutinise generative AI, and end users can make informed decisions.
For scientific background on this project, check out these peer-reviewed papers:
- Liesenfeld, A., & Dingemanse, M. (2024). Rethinking open source generative AI: open-washing and the EU AI Act. Proceedings of the 2024 ACM Conference on Fairness, Accountability, and Transparency (FAccT ’24). doi: 10.1145/3630106.3659005
- Liesenfeld, A., Lopez, A., & Dingemanse, M. (2023). Opening up ChatGPT: tracking openness, transparency, and accountability in instruction-tuned text generators. CUI ’23: Proceedings of the 5th International Conference on Conversational User Interfaces. doi: 10.1145/3571884.3604316
- Solaiman, I. (2023). The Gradient of Generative AI Release: Methods and Considerations. Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, 111–122. doi: 10.1145/3593013.3593981