search
HomeTechnology peripheralsAILeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

The emergence of GPT-4 has made many people worried about their scientific research, and even joked that NLP does not exist.

# Instead of worrying, it is better to use it in scientific research, simply "change the paper method".

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

Kareem Carr, a PhD in biostatistics from Harvard University, said that he has used large language model tools such as GPT-4 Conducted academic research.

# He said that these tools are very powerful, but they also have some very painful pitfalls.

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

His tweets about LLM usage advice even earned LeCun a recommendation.

#Let’s take a look at how Kareem Carr uses AI tools to conduct scientific research.

First Principle: Don’t look for LLM for content you cannot verify

At the beginning, Carr gave the first and most important rule Principle:

#Never ask a large language model (LLM) for information that you cannot verify yourself, or ask it to perform a task that you cannot verify that it has been completed correctly.

The only exception is if it is not a critical task, such as asking the LLM for apartment decorating ideas.

#"Using best practices in literature review, summarize research on breast cancer research over the past 10 years." This is a poor request because you cannot directly verify that it summarizes the literature correctly.

# Instead ask, “Give me a list of the top review articles on breast cancer research in the past 10 years.”

# Such prompts can not only verify the source, but also verify the reliability yourself.

Tips for writing "prompts"

It's very easy to ask LLM to write code for you or find relevant information, but the quality of the output content may be affected. There is a big difference. Here are some things you can do to improve quality:

Set context:

•Explicitly tell the LLM what information should be used

•Use terminology and symbols to orient the LLM towards the correct contextual information

If you have an idea on how to handle the request, please tell LLM the specific method to use. For example, "solve this inequality" should be changed to "use the Cauchy-Schwarz theorem to solve this inequality, and then apply the complete square."

#Be aware that these language models are much more linguistically complex than you think, and even very vague hints will be helpful.

Be more specific:

This is not a Google search , so don't worry if there's a site discussing your exact problem.

"How to solve the simultaneous equations of quadratic terms?" This prompt is not clear. You should ask: "Solve x=(1/2 )(a b) and y=(1/3)(a^2 ab b^2) A system of equations about a and b."

Define the output format:

Leverage the flexibility of LLMs to format the output to best suit Your way, such as:

• Code

• Math formula

• Articles

##• Tutorials

• A Brief Guide

#You can even ask for the code that generates the following, including tables, plots, charts.

#Although you get what LLM outputs, this is only the beginning. Because you need to verify the output content. This includes:

• Finding inconsistencies

• Searching tool output via Google Terminology of content, obtaining supportable sources

• When possible, write code to test yourself

The reason for self-verification is that LLMs often make strange mistakes that are inconsistent with their seeming professionalism. For example, the LLM may mention a very advanced mathematical concept but be confused about a simple algebra problem.

Ask one more time:

Large-scale language model generation The content is random. Sometimes, creating a new window and asking your question again may provide you with a better answer.

#In addition, use multiple LLM tools. Kareem Carr currently uses Bing AI, GPT-4, GPT-3.5 and Bard AI in scientific research according to his own needs. However, they each have their own advantages and disadvantages.

Quote Productivity

Quote

According to Carr's experience, it is best to ask the same mathematical questions to both GPT-4 and Bard AI at the same time to get different perspectives. Bing AI works on web searches. GPT-4 is much smarter than GPT-3.5, but currently OpenAI is limited to 25 messages in 3 hours, making it more difficult to access.

#As for the issue of citation, citing references is a particularly weak point of LLM. Sometimes the references LLM gives you exist, sometimes they don't.

Previously, a netizen encountered the same problem. He said that he asked ChatGPT to provide reference materials involving the mathematical properties of lists, but ChatGPT generated an error message that did not follow. Non-existent references are what everyone calls the "illusion" problem.

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

However, Kareem Carr points out that false quotes are not completely useless.

#In his experience, the words in fabricated references are often related to real terms, as well as to researchers in related fields. So Googling these terms often gets you closer to the information you're looking for.

#In addition, Bing is also a good choice when searching for sources.

Productivity

There are many unrealistic sayings about LLM improving productivity, such as "LLM It can increase your productivity by 10 times or even 100 times."

In Carr’s experience, this kind of acceleration only makes sense if no work is double-checked, which is true for someone who is an academic. is irresponsible.

However, LLM has greatly improved Kareem Carr’s academic workflow, including:

- Prototype idea design - Identify dead ideas - Speed ​​up tedious data reformatting tasks - Learn new programming languages, packages and concepts - Google search

With the current LLM, Carr said he spends less time on what to do next. LLM can help him advance vague, or incomplete ideas into complete solutions.

# Additionally, LLM reduced the amount of time Carr spent on side projects unrelated to his primary goals.

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

I found that I got into a flow state and I was able to keep going. This means I can work longer hours without burning out.

Final word of advice: Be careful not to get sucked into a side hustle. The sudden increase in productivity from these tools can be intoxicating and potentially distracting for individuals.

## Regarding the experience of ChatGPT, Carr once posted a post on LinkedIn to share his feelings after using ChatGPT:

As a data scientist, I have been experimenting with OpenAI’s ChatGPT for a few weeks. It's not as good as people think.

#Despite the initial disappointment, my feeling is that a system like ChatGPT can add tremendous value to the standard data analysis workflow.

At this point it's not obvious where this value lies. ChatGPT can easily get some details wrong on simple things, and it simply can't solve problems that require multiple inference steps.

#The main question for each new task in the future remains whether it is easier to evaluate and improve ChatGPT's solution attempts, or to start from scratch.

#I did find that even a poor solution to ChatGPT tended to activate relevant parts of my brain that starting from scratch did not.

#Like they always say it’s always easier to criticize a plan than to come up with one yourself.

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

Netizens need to verify the content output by AI, saying that in most cases, artificial intelligence The accuracy rate is about 90%. But the remaining 10% of mistakes can be fatal.

Carr joked, if it was 100%, then I wouldn’t have a job.

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

So, why does ChatGPT generate false references?

It is worth noting that ChatGPT uses a statistical model to guess the next word, sentence and paragraph based on probability to match the context provided by the user.

Because the source data of the language model is very large, it needs to be "compressed", which causes the final statistical model to lose accuracy.

This means that even if there are true statements in the original data, the "distortion" of the model will create a "fuzziness" that causes the model to produce the most " "Plausible" statement.

#In short, this model does not have the ability to evaluate whether the output it produces is equivalent to a true statement.

In addition, the model is created based on crawling or crawling public network data collected through the public welfare organization "Common Crawl" and similar sources. The data is as of 21 years.

#Since data on the public Internet is largely unfiltered, this data may contain a large amount of erroneous information.

LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow

Recently, an analysis by NewsGuard found that GPT-4 is actually more likely to generate error messages than GPT-3.5 , and the persuasiveness in the reply is more detailed and convincing.

In January, NewsGuard first tested GPT-3.5 and found that it generated 80 out of 100 fake news narratives. A subsequent test of GPT-4 in March found that GPT-4 responded falsely and misleadingly to all 100 false narratives.

#It can be seen that source verification and testing are required during the use of LLM tools.

The above is the detailed content of LeCun highly recommends! Harvard doctor shares how to use GPT-4 for scientific research, down to every workflow. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:51CTO.COM. If there is any infringement, please contact admin@php.cn delete
What is Graph of Thought in Prompt EngineeringWhat is Graph of Thought in Prompt EngineeringApr 13, 2025 am 11:53 AM

Introduction In prompt engineering, “Graph of Thought” refers to a novel approach that uses graph theory to structure and guide AI’s reasoning process. Unlike traditional methods, which often involve linear s

Optimize Your Organisation's Email Marketing with GenAI AgentsOptimize Your Organisation's Email Marketing with GenAI AgentsApr 13, 2025 am 11:44 AM

Introduction Congratulations! You run a successful business. Through your web pages, social media campaigns, webinars, conferences, free resources, and other sources, you collect 5000 email IDs daily. The next obvious step is

Real-Time App Performance Monitoring with Apache PinotReal-Time App Performance Monitoring with Apache PinotApr 13, 2025 am 11:40 AM

Introduction In today’s fast-paced software development environment, ensuring optimal application performance is crucial. Monitoring real-time metrics such as response times, error rates, and resource utilization can help main

ChatGPT Hits 1 Billion Users? 'Doubled In Just Weeks' Says OpenAI CEOChatGPT Hits 1 Billion Users? 'Doubled In Just Weeks' Says OpenAI CEOApr 13, 2025 am 11:23 AM

“How many users do you have?” he prodded. “I think the last time we said was 500 million weekly actives, and it is growing very rapidly,” replied Altman. “You told me that it like doubled in just a few weeks,” Anderson continued. “I said that priv

Pixtral-12B: Mistral AI's First Multimodal Model - Analytics VidhyaPixtral-12B: Mistral AI's First Multimodal Model - Analytics VidhyaApr 13, 2025 am 11:20 AM

Introduction Mistral has released its very first multimodal model, namely the Pixtral-12B-2409. This model is built upon Mistral’s 12 Billion parameter, Nemo 12B. What sets this model apart? It can now take both images and tex

Agentic Frameworks for Generative AI Applications - Analytics VidhyaAgentic Frameworks for Generative AI Applications - Analytics VidhyaApr 13, 2025 am 11:13 AM

Imagine having an AI-powered assistant that not only responds to your queries but also autonomously gathers information, executes tasks, and even handles multiple types of data—text, images, and code. Sounds futuristic? In this a

Applications of Generative AI in the Financial SectorApplications of Generative AI in the Financial SectorApr 13, 2025 am 11:12 AM

Introduction The finance industry is the cornerstone of any country’s development, as it drives economic growth by facilitating efficient transactions and credit availability. The ease with which transactions occur and credit

Guide to Online Learning and Passive-Aggressive AlgorithmsGuide to Online Learning and Passive-Aggressive AlgorithmsApr 13, 2025 am 11:09 AM

Introduction Data is being generated at an unprecedented rate from sources such as social media, financial transactions, and e-commerce platforms. Handling this continuous stream of information is a challenge, but it offers an

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
3 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Best Graphic Settings
3 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. How to Fix Audio if You Can't Hear Anyone
3 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
WWE 2K25: How To Unlock Everything In MyRise
4 weeks agoBy尊渡假赌尊渡假赌尊渡假赌

Hot Tools

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

mPDF

mPDF

mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

DVWA

DVWA

Damn Vulnerable Web App (DVWA) is a PHP/MySQL web application that is very vulnerable. Its main goals are to be an aid for security professionals to test their skills and tools in a legal environment, to help web developers better understand the process of securing web applications, and to help teachers/students teach/learn in a classroom environment Web application security. The goal of DVWA is to practice some of the most common web vulnerabilities through a simple and straightforward interface, with varying degrees of difficulty. Please note that this software

Dreamweaver Mac version

Dreamweaver Mac version

Visual web development tools

SecLists

SecLists

SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.