


GPT4All is a chatbot trained based on a large amount of clean assistant data (including code, stories and conversations). The data includes ~800k pieces of GPT-3.5-Turbo generated data. It is completed based on LLaMa and can be used in M1 Mac, Windows and other environments. Can run. Perhaps as its name suggests, the time has come when everyone can use personal GPT.
Since OpenAI released ChatGPT, chatbots have become increasingly popular in recent months.
Although ChatGPT is powerful, it is almost impossible for OpenAI to open source it. Many people are working on open source, such as LLaMA, which was open sourced by Meta some time ago. It is a general term for a series of models with parameter quantities ranging from 7 billion to 65 billion. Among them, the 13 billion parameter LLaMA model can outperform the 175 billion parameter GPT-3 "on most benchmarks".
The open source of LLaMA has benefited many researchers. For example, Stanford added instruct tuning to LLaMA and trained a new 7 billion parameter model called Alpaca (based on LLaMA 7B ). The results show that the performance of Alpaca, a lightweight model with only 7B parameters, is comparable to very large-scale language models such as GPT-3.5.
For another example, the model we are going to introduce next, GPT4All, is also a new 7B language model based on LLaMA. Two days after the project was launched, the number of Stars has exceeded 7.8k.
Project address: https://github.com/nomic-ai/gpt4all
To put it simply, GPT4All is 800k in GPT-3.5-Turbo Training is performed on pieces of data, including text questions, story descriptions, multiple rounds of dialogue and code.
According to the project display, M1 can run in Mac, Windows and other environments.
Let’s take a look at the effect first. As shown in the figure below, users can communicate with GPT4All without any barriers, such as asking the model: "Can I run a large language model on a laptop?" GPT4All's answer is: "Yes, you can use a laptop to train and test neural networks Or machine learning models for other natural languages (such as English or Chinese). Importantly, you need enough available memory (RAM) to accommodate the size of these models..."
Next, if you don't know You can continue to ask GPT4All how much memory is needed, and it will give an answer. Judging from the results, GPT4All’s ability to conduct multiple rounds of dialogue is still very strong.
Real-time sampling on M1 Mac
Some people call this research "game-changing" Rules, with the blessing of GPT4All, you can now run GPT locally on MacBook."
Similar to GPT-4, GPT4All also provides a "technical Report".
Technical report address: https://s3.amazonaws.com/static.nomic.ai/gpt4all/2023_GPT4All_Technical_Report.pdf
This preliminary The technical report briefly describes the construction details of GPT4All. The researchers disclosed the collected data, data wrangling procedures, training code, and final model weights to promote open research and reproducibility. They also released a quantized 4-bit version of the model, which means almost Anyone can run the model on a CPU.
Next, let’s take a look at what’s written in this report.
GPT4All Technical Report
1. Data collection and sorting
During the period from March 20, 2023 to March 26, 2023, the research Researchers used the GPT-3.5-Turbo OpenAI API to collect approximately 1 million pairs of prompt responses.
First, the researchers collected different question/prompt samples by utilizing three publicly available data sets:
- Unified chip2 subset of LAION OIG
- A random subsample set of Stackoverflow Questions Coding questions
- Bigscience/P3 subsample set for instruction tuning
Referring to the Stanford University Alpaca project (Taori et al., 2023), researchers have paid a lot of attention to data preparation and organization. After collecting the initial dataset of prompt-generated pairs, they loaded the data into Atlas to organize and clean it, removing any samples in which GPT-3.5-Turbo failed to respond to prompts and produced malformed output. This reduces the total number of samples to 806199 high-quality prompt-generated pairs. Next, we removed the entire Bigscience/P3 subset from the final training dataset because its output diversity was very low. P3 contains many homogeneous prompts that generate short and homogeneous responses from GPT-3.5-Turbo.
This elimination method resulted in a final subset of 437,605 prompt-generated pairs, as shown in Figure 2.
Model training
The researchers combined several models in an instance of LLaMA 7B (Touvron et al., 2023) Make fine adjustments. Their original public release-related model was trained with LoRA (Hu et al., 2021) on 437,605 post-processed examples for 4 epochs. Detailed model hyperparameters and training code can be found in the relevant resource library and model training logs.
Reproducibility
The researchers released all data (including unused P3 generations), training code and model weights for the community to reproduce. Interested researchers can find the latest data, training details, and checkpoints in the Git repository.
Cost
It took the researchers about four days to create these models, and the GPU cost was $800 (rented from Lambda Labs and Paperspace, including several failed training), in addition to the $500 OpenAI API fee.
The final released model gpt4all-lora can be trained in approximately 8 hours on Lambda Labs’ DGX A100 8x 80GB, at a total cost of $100.
This model can be run on an ordinary laptop. As a netizen said, "There is no cost except the electricity bill."
Evaluation
The researchers conducted a preliminary evaluation of the model using human evaluation data in the SelfInstruct paper (Wang et al., 2022). The report also compares the ground truth perplexity of this model with the best known public alpaca-lora model (contributed by huggingface user chainyo). They found that all models had very large perplexities on a small number of tasks and reported perplexities up to 100. Models fine-tuned on this collected dataset showed lower perplexity in Self-Instruct evaluation compared to Alpaca. The researchers say this evaluation is not exhaustive and there is still room for further evaluation - they welcome readers to run the model on a local CPU (documentation available on Github) and get a qualitative sense of its capabilities.
Finally, it is important to note that the authors publish data and training details in the hope that it will accelerate open LLM research, especially in the areas of alignment and interpretability. GPT4All model weights and data are for research purposes only and are licensed against any commercial use. GPT4All is based on LLaMA, which has a non-commercial license. Assistant data was collected from OpenAI's GPT-3.5-Turbo, whose terms of use prohibit the development of models that compete commercially with OpenAI.
The above is the detailed content of A replacement for ChatGPT that can be run on a laptop is here, with a full technical report attached. For more information, please follow other related articles on the PHP Chinese website!

把逻辑地址转换为内存物理地址的过程称作“重定位”。程序发出的逻辑地址并不是数据或指令的真实存放地,但可以对这个地址进行转换来获得真实存放地的物理地址,相当于重新定位一次。

根据计算机网络覆盖范围,可将计算机网络分为三类:1、局域网(LAN),是一种在小区域内使用的,由多台计算机组成的网络,覆盖范围通常局限在10千米范围之内;2、广域网(WAN),是一种远程网,涉及长距离的通信,覆盖范围可以是个国家或多个国家,甚至整个世界;3、城域网(MAN),其网络覆盖范围通常可以延伸到整个城市,借助通信光纤将多个局域网联通公用城市网络形成大型网络。

微机的字长是4个字节意味着:在CPU中整体传输和处理的二进制数为32位。因为一个字节是8位长(字长),所以4个字节就是32位了,也就是说cpu中能够同时处理32位的二进制数据。在计算机领域,字是用来表示一次性处理事务的固定长度;一个字的位数,即字长,是计算机一次可处理的二进制数字的数目。

微型计算机的性能主要取决于CPU(中央处理器)的性能。CPU是计算机系统的运算和控制核心,是对计算机的所有硬件资源(如存储器、输入输出单元) 进行控制调配、执行通用运算的核心硬件单元;中央处理器(CPU)的性能对计算机性能起决定性作用。

微型计算机的运算器控制器及内存储器的总称是“主机”。在微型计算机中,运算器、控制器、存储器三个部分是信息加工、处理的主要部件;运算器和控制器总称为CPU(中央处理单元),而CPU与内存储器又总称为主机,这是计算机系统中最核心的硬件部分。

计算机系统的内部总线主要可以分为5类:1、数据总线,在CPU与RAM之间来回传送需要处理或是需要储存的数据;2、地址总线,用来指定在RAM之中储存的数据的地址;3、控制总线,将微处理器控制单元的信号,传送到周边设备;4、扩展总线,是外部设备和计算机主机进行数据通信的总线,例如ISA总线,PCI总线;5、局部总线,取代更高速数据传输的扩展总线。

MODEM的功能为“实现模拟信号与数字信号之间的相互转换”。MODEM的中文名为“调制解调器”,它可以在发送端通过调制将数字信号转换成通信线路上传输的模拟信号,在接收端通过解调再将模拟信号转换为数字信号。

世界上第一台计算机的电子元器件是“电子真空管”。世界上第一台计算机是“阿塔纳索夫-贝瑞计算机”,通常简称ABC计算机,采用电子真空管作为电子元件;该计算机电路系统中装有300个电子真空管执行数字计算与逻辑运算,机器使用电容器来进行数值存储,数据输入采用打孔读卡方法,还采用了二进位制。


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

SecLists
SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.

MinGW - Minimalist GNU for Windows
This project is in the process of being migrated to osdn.net/projects/mingw, you can continue to follow us there. MinGW: A native Windows port of the GNU Compiler Collection (GCC), freely distributable import libraries and header files for building native Windows applications; includes extensions to the MSVC runtime to support C99 functionality. All MinGW software can run on 64-bit Windows platforms.

SublimeText3 Mac version
God-level code editing software (SublimeText3)

SAP NetWeaver Server Adapter for Eclipse
Integrate Eclipse with SAP NetWeaver application server.

Zend Studio 13.0.1
Powerful PHP integrated development environment
