Stanford University publishes model 'Alpaca 7B' that can build chat AI comparable to GPT-3.5, open source and inexpensively reproducible



A research team at Stanford University has released a large-scale language model `` Alpaca 7B '' that achieves better command followability with open source, which is a fine-tuned 7B model of Meta's large-scale language model ``

LLaMA ''. Alpaca makes it possible to reproduce GPT-3.5-like behavior easily and cheaply in a much smaller environment.

Alpaca: A Strong Open-Source Instruction-Following Model - Stanford CRFM
https://crfm.stanford.edu/2023/03/13/alpaca.html



Stanford Alpaca, and the acceleration of on-device large language model development

https://simonwillison.net/2023/Mar/13/alpaca/



According to Stanford University, while language models such as GPT-3.5 (text-davinci-003), Chat GPT, Claude, and Bing Chat are becoming more and more powerful, there are still many flaws left. . Maximum progress toward solving this problem required the involvement of academia, but the lack of an open-source model with comparable functionality to closed models such as text-davinci-003 meant that academia was unable to follow orders. It seems that it was difficult to conduct model research.

Therefore, Stanford University made fine adjustments based on the 7B model of Meta's large-scale language model 'LLaMA', which was trained only with a public dataset, and developed and released a language model called 'Alpaca'. Fine adjustment took 3 hours using 8

NVIDIA A100 with 80 GB memory. This is available for less than $ 100 (about 13,400 yen) at many cloud computing providers.

Alpaca exhibits many similar behaviors to text-davinci-003, but is surprisingly small, easy, and cheap to reproduce.

In fact, a web demo is available on the following site and you can try it.

Alpaca
https://alpaca-ai-custom5.ngrok.io/



Please note that the use of Alpaca is limited to academic purposes and commercial use is prohibited. This is because the base LLaMA is a non-commercial license, the instruction data is based on text-davinci-003, which prohibits the development of models that compete with OpenAI, and the safety measures are not sufficient to spread it to the general public. It seems that it is due to three reasons that it can not be done.

in Note, Posted by logc_nt