European Open Source AI Index
DatabaseNewsGuidesAboutContribute

StableVicuna

by CarperAI

StableVicuna-13B is a Vicuna-13B v0 model fine-tuned using reinforcement learning from human feedback (RLHF) via Proximal Policy Optimization (PPO) on various conversational and instructional datasets
Text
Full
https://huggingface.co/CarperAI/stable-vicuna-13b-delta
LLaMA-13B
StableVicuna-13B
Meta Community License
CarperAI, an AI research team spun out of EleutherAI.
https://carper.ai
April 2023
Availability
Base Model Data
Based on LLaMA whose pretraining data has nowhere been disclosed or documented.
https://github.com/facebookresearch/llama/blob/main/MODEL_CARD.md#training-dataset
End User Model Data
From the documentation 'The reward model used during RLHF was also trained on OpenAssistant Conversations Dataset (OASST1) along with two other datasets Anthropic HH-RLHF, a dataset of preferences about AI assistant helpfulness and harmlessness; and Stanford Human Preferences Dataset a dataset of 385K collective human preferences over responses to questions/instructions in 18 different subject areas, from cooking to legal advice.'
https://huggingface.co/CarperAI/stable-vicuna-13b-delta
Base Model Weights
Based on LLaMA weights, which are not openly available though a leaked versions is in wide circulation.
End User Model Weights
Model not functional out of the box as weights require a delta computation. From the docs 'StableVicuna-13B cannot be used from the CarperAI/stable-vicuna-13b-delta weights alone. To obtain the correct model, one must add back the difference between LLaMA 13B and CarperAI/stable-vicuna-13b-delta weights.'
https://huggingface.co/CarperAI/stable-vicuna-13b-delta#apply-delta-weights
Training Code
Training code not made public, only code for applying the delta.
Documentation
Code Documentation
Code is minimally documented and deployment requires non-trivial configuration, e.g. 'StableVicuna-13B cannot be used from the CarperAI/stable-vicuna-13b-delta weights alone. To obtain the correct model, one must add back the difference between LLaMA 13B and CarperAI/stable-vicuna-13b-delta weights.'
https://huggingface.co/CarperAI/stable-vicuna-13b-delta/tree/main
Hardware Architecture
Architecture is described at a high level in scattered places, but there is no clear and exhaustive overview.
Preprint
No preprint available.
Paper
No peer-reviewed paper found.
Modelcard
Model card provides some information but is not fully worked out as recommended in model card literature.
https://huggingface.co/lmsys/vicuna-13b-delta-v0
Datasheet
No datasheet found
Access
Licenses
CC-BY-NC-SA-4.0. License for LLaMA is more murky, hence partial. As they say 'License for the base LLaMA model's weights is Meta's non-commercial bespoke license.'
https://huggingface.co/CarperAI/stable-vicuna-13b-delta
Last updated 13 March 2026
Is this information not up to date?
Contribute here ->
Supported by the Centre for Language Studies and the Dutch Research Council. Website design & development © 2024 by BSTN. This version of the index generated 09 April 2026, website content last updated 11 March 2026.