The idea behind this is to create an affordable alternative to OpenAI’s GPT-3.5 (text-davinci-003) that can still perform well on tasks like email writing, social media management, and productivity tools.
The training recipe for Alpaca involves using Meta’s LLaMA models as the starting point and fine-tuning them with 52K unique instructions generated from OpenAI’s text-davinci-003. This process takes about three hours on eight NVIDIA A100 GPUs, which costs less than $100 on most cloud compute providers.
The preliminary evaluation of Alpaca shows that it performs similarly to text-davinci-003 in terms of accuracy and style but with a smaller model size and lower cost. However, like any language model, Alpaca also exhibits common deficiencies such as hallucination (spreading misinformation) and toxicity (producing negative or hateful content).
To mitigate these risks, the authors of this paper have implemented a content filter using OpenAI’s moderation API to prevent the dissemination of harmful content. They also watermark all model outputs to make it easier for researchers to detect whether an output comes from Alpaca 7B or not.
Overall, the release of Alpaca is intended only for academic research and any commercial use is prohibited. The authors believe that this will enable more people (including bad actors) to create models that could cause harm but they also hope that it will incentivize swift defensive action from the academic community. By installing these mitigations, they goal to advance best practices for responsible deployment of foundation models and ultimately develop community norms around their use.
In terms of future directions, the authors plan to evaluate Alpaca more rigorously using methods like HELM (Holistic Evaluation of Language Models) and study how capabilities arise from the training recipe. They also hope to better understand what properties of a base model are needed for instruction-following tasks and whether there are alternatives to using self-instruct on text-davinci-003.
So if you’re looking for an affordable alternative to OpenAI’s GPT-3.5 that can still perform well on instruction-following tasks, Stanford Alpaca might be worth checking out!