Data of Meta's large-scale language model 'LLaMA-65B' leaked on 4chan



The large-scale language model ``

LLaMA '' announced by Meta in February 2023 has performance comparable to GPT-3 and is light enough to operate even with a single GPU, so ChatGPT can be used even in consumer-level hardware environments. It is expected to be able to operate AI like It has become a hot topic that such LLaMA data leaked.

Facebook LLAMA is being openly distributed via torrents | Hacker News
https://news.ycombinator.com/item?id=35007978

LLaMA is a large-scale language model developed by Meta AI Research, Meta's AI research organization. In order to operate conventional large-scale language models such as OpenAI's ChatGPT and DeepMind's Chinchilla , it was necessary to use multiple accelerators optimized for AI, but LLaMA can operate sufficiently even with a single GPU. , the advantage is that the number of parameters that indicate the scale of the model is overwhelmingly small. At the time of writing the article, part of the model data is published on GitHub, and if you contact Meta AI Research, you can download the `` weights '' learned by the neural network separately.

Meta announces large-scale language model 'LLaMA', can operate with a single GPU while having performance comparable to GPT-3 - GIGAZINE



However, on March 3, 2023, a user named ``llamanon !!T2UdrWkLSWB'' suddenly downloaded the ``weight'' data of LLaMA-65B (65 billion parameters) in a thread about AI chatbots on the online bulletin board site 4chan. I published a torrent file and a magnet link that I can.



There is also a pull request to add the magnet link published on 4chan to LLaMA's repository on GitHub.

Save bandwidth by using a torrent to distribute more efficiently by ChristopherKing42 · Pull Request #73 · facebookresearch/llama · GitHub
https://github.com/facebookresearch/llama/pull/73/files



Furthermore, based on this leaked LLaMA weight data, LLaMA 7B (7 billion parameters), 13B (13 billion parameters), 30B (30 billion parameters), and 65B weight data can be downloaded at 40MB/s. A downloader is available on GitHub.

GitHub - shawwn/llama-dl: High-speed download of LLaMA, Facebook's 65B parameter GPT model
https://github.com/shawwn/llama-dl



Sean Presser , who released the downloader, said, ``Some people have already claimed that the weight data of LLaMA has been leaked. However, the GPT-2 1.5B (1.5 billion parameters) model leak In fact, the great appeal of GPT-2 was the driving force behind me getting serious about machine learning in 2019. Four years later, in 2023, So, it's clear that no one cares about the leak model of GPT-2 anymore, and there was no widespread social harm, and LLaMA will be the same.'

in Software,   Web Service, Posted by log1i_yk