Home >Technology peripherals >AI >Attention-free large model Eagle7B: Based on RWKV, inference cost is reduced by 10-100 times

Attention-free large model Eagle7B: Based on RWKV, inference cost is reduced by 10-100 times

WBOY
WBOYforward
2024-02-01 14:39:461186browse

Attention-free large model Eagle7B: Based on RWKV, the inference cost is reduced by 10-100 times

In the AI ​​​​track, small models have attracted much attention recently. Compared with models with hundreds of billions of parameters, Model. For example, the Mistral-7B model released by the French AI startup outperformed Llama 2 by 13B in every benchmark, and outperformed Llama 1 by 34B in code, math, and inference.

Compared with large models, small models have many advantages, such as low computing power requirements and the ability to run on the device side.

Recently, a new language model has emerged, namely the 7.52B parameter Eagle 7B, from the open source non-profit organization RWKV, which has the following characteristics:

无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍


  • Built based on RWKV-v5 architecture, which has lower reasoning cost (RWKV is a linear transformer, reducing reasoning cost 10-100 times more);
  • is trained on more than 100 languages ​​and 1.1 trillion tokens;
  • is trained on multi-language benchmarks In the test, it is better than all 7B class models;
  • In the English evaluation, the performance of Eagle 7B is close to Falcon (1.5T), LLaMA2 (2T), and Mistral;
  • Compared with MPT-7B (1T) in English review;
  • Transformer without attention.

无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍

##Eagle 7B is built based on the RWKV-v5 architecture. RWKV (Receptance Weighted Key Value) is a novel architecture that combines the advantages of RNN and Transformer and avoids their shortcomings. It is very well designed and can alleviate the memory and expansion bottlenecks of Transformer and achieve more effective linear expansion. At the same time, RWKV also retains some of the properties that have made Transformer dominant in the field.

Currently RWKV has been iterated to the sixth generation RWKV-6, with performance and size similar to Transformer. Future researchers can use this architecture to create more efficient models.

For more information about RWKV, you can refer to "Reshaping RNN in the Transformer era, RWKV expands the non-Transformer architecture to tens of billions of parameters."

It is worth mentioning that RWKV-v5 Eagle 7B can be used for personal or commercial use without restrictions.

Test results on 23 languages

The performance of different models on multiple languages ​​is as follows, Test benchmarks include xLAMBDA, xStoryCloze, xWinograd, and xCopa.

无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍


无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍

##23 languages ​​in total

These benchmarks incorporate mostly common sense reasoning and show the huge leap in multi-language performance of the RWKV architecture from v4 to v5. However, due to the lack of multilingual benchmarks, the study can only test its ability in 23 more commonly used languages, and the ability in the remaining 75 or more languages ​​is still unknown.

Performance in English

The performance of different models in English is judged through 12 benchmarks, including Common sense reasoning and world knowledge.

无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍

From the results we can once again see RWKV’s huge leap from v4 to v5 architecture. v4 previously lost to the 1T token MPT-7b, but v5 has begun to catch up in benchmark tests. In some cases (even in some benchmark tests LAMBADA, StoryCloze16, WinoGrande, HeadQA_en, Sciq) it can surpass Falcon , or even llama2.

Additionally, v5 performance starts to align with expected Transformer performance levels given the approximate token training statistics.

Previously, Mistral-7B used a training method of 2-7 trillion Tokens to maintain its lead in the 7B scale model. The study hopes to close this gap so that RWKV-v5 Eagle 7B surpasses llama2 performance and reaches the level of Mistral.

The following figure shows that RWKV-v5 Eagle 7B's checkpoints near 300 billion token points show similar performance to pythia-6.9b:

无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍

This is consistent with previous experiments (pile-based) on the RWKV-v4 architecture, where linear transformers like RWKV are similar in performance level to transformers and have the same number of tokens. train.

无注意力大模型Eagle7B:基于RWKV,推理成本降低10-100 倍

# Predictably, the emergence of this model marks the arrival of the strongest linear transformer (in terms of evaluation benchmarks) to date.

The above is the detailed content of Attention-free large model Eagle7B: Based on RWKV, inference cost is reduced by 10-100 times. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:51cto.com. If there is any infringement, please contact admin@php.cn delete