Home >Technology peripherals >AI >Choose GPT-3.5 or Jordan Llama 2 and other open source models? After comprehensive comparison, the answer is

Choose GPT-3.5 or Jordan Llama 2 and other open source models? After comprehensive comparison, the answer is

WBOY
WBOYforward
2023-10-16 18:45:05577browse

By comparing the parameters of GPT-3.5 and Llama 2 on different tasks, we can know under what circumstances choose GPT-3.5 and under what circumstances choose Llama 2 or other models.

Apparently, torqueing GPT-3.5 is very expensive. This paper experimentally verifies whether a manual torque model can approach the performance of GPT-3.5 at a fraction of the cost of GPT-3.5. Interestingly, the paper did.

Comparing the results on SQL tasks and function representation tasks, the paper found that:

  • GPT-3.5 performed well in two data sets (a subset of the Spider data set and Viggo Function representation data set) are slightly better than Code Llama 34B through Lora.

  • The training cost of GPT-3.5 is 4-6 times higher, and the deployment cost is also higher.

One of the conclusions of this experiment is that GPT-3.5 is suitable for initial verification work, but after that, a model like Llama 2 may be the best choice. To summarize briefly:

  • If you want validation to be the right way to solve a specific task/dataset, or want a fully managed environment, then adjust GPT-3.5.

  • If you want to save money, get maximum performance from your data set, have more flexibility in training and deploying infrastructure, and want or retain some data, then Just consume an open source model like Llama 2.

Next let’s see how the paper is implemented.

The following figure shows the performance of Code Llama 34B and GPT-3.5 trained to convergence on SQL tasks and function representation tasks. The results show that GPT-3.5 achieves better accuracy on both tasks.

选择GPT-3.5、还是乔丹Llama 2等开源模型?综合比较后答案有了

In terms of hardware usage, the experiment used an A40 GPU, which is approximately US$0.475.

选择GPT-3.5、还是乔丹Llama 2等开源模型?综合比较后答案有了

#In addition, the experiment enumerates two data sets that are very suitable for terrible experiments. The Spider data set is a subset of the Viggo function representing the data set.

In order to make a fair comparison with the GPT-3.5 model, experiments were performed on Llama with minimal hyperparameters.

Two key choices for this article’s experiments are to use Code Llama 34B and Lora parameters instead of full-parameter parameters.

The rules for Lora hyperparameter configuration were followed to a large extent in the experiment. The Lora load is as follows:

选择GPT-3.5、还是乔丹Llama 2等开源模型?综合比较后答案有了

SQL prompt examples are as follows:

选择GPT-3.5、还是乔丹Llama 2等开源模型?综合比较后答案有了

# SQL Reminder part of the display, please check the original blog

This experiment does not use a complete SPIDER dataset, specific The form is as follows

department : Department_ID [ INT ] primary_key Name [ TEXT ] Creation [ TEXT ] Ranking [ INT ] Budget_in_Billions [ INT ] Num_Employees [ INT ] head : head_ID [ INT ] primary_key name [ TEXT ] born_state [ TEXT ] age [ INT ] management : department_ID [ INT ] primary_key management.department_ID = department.Department_ID head_ID [ INT ] management.head_ID = head.head_ID temporary_acting [ TEXT ]

The experiment chooses to use the intersection of sql-create-context data set and Spider data set. The context provided for the model is a SQL creation command as follows:

CREATE TABLE table_name_12 (class VARCHAR, frequency_mhz VARCHAR, city_of_license VARCHAR)

The code and data address of the SQL task: https://github.com/samlhuillier/spider-sql- finetune

The example of the function representation prompt is as follows: 选择GPT-3.5、还是乔丹Llama 2等开源模型?综合比较后答案有了

#The output is as follows:
verify_attribute(name[Little Big Adventure], rating[average], has_multiplayer[no], platforms[PlayStation])

During the evaluation phase, the two experiments quickly converged:

选择GPT-3.5、还是乔丹Llama 2等开源模型?综合比较后答案有了

The function represents the task code and data address: https://github.com/samlhuillier/viggo-finetune

For more information, please check the original blog.

Original link:

https://ragntune.com/blog/gpt3.5-vs-llama2 -finetuning?continueFlag=11fc7786e20d498fc4daa79c5923e198

###

The above is the detailed content of Choose GPT-3.5 or Jordan Llama 2 and other open source models? After comprehensive comparison, the answer is. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:jiqizhixin.com. If there is any infringement, please contact admin@php.cn delete