Announcement of chatbot 'GPT4ALL' with 7 billion parameters that can be run on notebook PCs for free



Nomic AI announced a chatbot `` GPT4ALL '' that can be executed on a notebook PC using data learned by GPT-3.5-Turbo and Meta's large-scale language model ``

LLaMA ''. All of the training datasets of question-response pairs used to train the various topics are also publicly available.

GitHub - nomic-ai/gpt4all: gpt4all: a chatbot trained on a massive collection of clean assistant data including code, stories and dialogue
https://github.com/nomic-ai/gpt4all



GPT4All: Training an Assistant-style Chatbot with Large Scale Data Distillation from GPT-3.5-Turbo
(PDF file)

https://s3.amazonaws.com/static.nomic.ai/gpt4all/2023_GPT4All_Technical_Report.pdf



GPT4All: Running an Open-source ChatGPT Clone on Your Laptop | by Maximilian Strauss | Mar, 2023 | Better Programming
https://betterprogramming.pub/gpt4all-running-an-open-source-chatgpt-clone-on-your-laptop-71ebe8600c71

Nomic AI first used GPT-3.5-Turbo to collect approximately 1 million question-answer pairs.



Based on the preceding commentary by Stanford University's Alpaca , Nomic AI decided to pay attention to data preparation and curation, and organized the collected pairs using a tool called Atlas . We have removed low-diversity questions so that the training data covers a wide range of topics. There are 437,605 pairs left after sorting.



And Nomic AI trains multiple fine-tuned models from an instance of LLaMA 7B. The model associated with the first publication was trained on LoRA. Comparing this model with the open-source 'Alpaca-LoRA' results in consistently low perplexity (lower is better).



GPT4ALL is an open-source lightweight clone of ChatGPT. Maximilian Strauss, author of the blog Better Programming, said, ``The appeal of GPT4ALL lies in the release of the 4-bit quantized model.'' This means that we run some parts of the model at a reduced precision, resulting in a more compact model that can run on consumer-level devices without dedicated hardware.

Below is a report that it worked on iMac M1 with 8GB of memory.



At hand, I confirmed that it works on the VAIO SX12 .



The model is licensed and available free of charge for research purposes only, but commercial use is prohibited. Also note that the terms of use of GPT-3.5-Turbo, which collected auxiliary data, prohibit the development of models that are commercially competitive with OpenAI.

in Software, Posted by logc_nt