search
HomeTechnology peripheralsAIIn addition to RAG, there are five ways to eliminate the illusion of large models

Produced by 51CTO Technology Stack (WeChat ID: blog51cto)

It is well known that LLM can produce hallucinations - that is, generate incorrect, misleading or meaningless information.

Interestingly, some people, such as OpenAI CEO Sam Altman, view the imagination of AI as creativity, while others believe that imagination may help make new scientific discoveries.

In most cases, however, it is crucial to provide the correct answer, and hallucinations are not a feature, but a flaw.

So, how to reduce the illusion of LLM? Long context? RAG? Fine-tuning?

In fact, long context LLMs are not foolproof, vector search RAG is not satisfactory, and fine-tuning comes with its own challenges and limitations.

The following are some advanced techniques that can be used to reduce the LLM illusion.

1. Advanced prompts

There is indeed a lot of discussion about whether using better or more advanced prompts can solve the hallucination problem of large language models (LLM).

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

Some people think that writing more detailed prompt words will not help solve the (hallucination) problem, but the co-founder of Google Brain (Google Brain) But people like Andrew Ng saw the potential. They proposed a new method that uses deep learning technology to generate prompt words to help people solve problems better. This method utilizes a large amount of data and powerful computing power to automatically generate prompt words related to the problem, thereby improving the efficiency of problem solving. Although the field

Ng believes that the inference capabilities of GPT-4 and other advanced models make them very good at interpreting complex prompt words with detailed instructions.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

“With multi-example learning, developers can give dozens, or even hundreds, of examples in a prompt word, which is better than less Learning by example is more effective,” he writes.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

In order to improve prompt words, many new developments are constantly emerging. For example, Anthropic released a new The “Prompt Generator” tool converts simple descriptions into high-level prompts optimized for large language models (LLMs). Through the Anthropic console, you can generate prompt words for production.

Recently, Marc Andreessen also said that with the right prompts, we can unlock the potential super genius in AI models. "Prompting techniques in different areas could unlock this potential super-genius," he added.

2.Meta AI’s Chain-of-Verification (CoVe)

Meta AI’s Chain-of-Verification (CoVe) is another technology. This approach reduces hallucinations in large language models (LLMs) by breaking down fact-checking into manageable steps, improving response accuracy, and aligning with human-driven fact-checking processes.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

CoVe involves generating an initial response, planning validation questions, answering those questions independently, and generating a final validated response. This approach significantly improves the accuracy of the model by systematically validating and correcting its output.

It improves performance in a variety of tasks such as list-based questions, closed-book question answering, and long-form text generation by reducing hallucinations and increasing factual correctness.

3. Knowledge Graph

RAG (Retrieval Enhanced Generation) is no longer limited to vector database matching. Many advanced RAG technologies have been introduced to significantly improve the retrieval effect.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

For example, integrate knowledge graphs (KGs) into RAG. By leveraging structured and interconnected data in knowledge graphs, the reasoning capabilities of current RAG systems can be greatly enhanced.

4.Raptor

Another technique is Raptor, which handles problems that span multiple documents by creating a higher level of abstraction. It is particularly useful when answering queries involving multiple document concepts.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

Approaches like Raptor fit well with long-context large language models (LLMs) because you can directly embed the entire document without chunking .

This method reduces the hallucination phenomenon by integrating an external retrieval mechanism with the transformer model. When a query is received, Raptor first retrieves relevant and verified information from external knowledge bases.

These retrieved data are then embedded into the context of the model along with the original query. By basing the model's responses on facts and relevant information, Raptor ensures that the content generated is both accurate and contextual.

5. Conformal Abstention

The paper "Relieving the Hallucination Phenomenon of Large Language Models through Conformal Abstention" introduces a method to determine the model by applying conformal prediction technology Methods to reduce hallucinations in large language models (LLMs) when responses should be avoided.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

By using self-consistency to evaluate the similarity of responses and leveraging conformal prediction for strict guarantees, this method ensures that the model only Only respond when you are confident in its accuracy.

This method effectively limits the incidence of hallucinations while maintaining a balanced withdrawal rate, which is especially beneficial for tasks that require long answers. It significantly improves the reliability of model output by avoiding erroneous or illogical responses.

6.RAG reduces the hallucination phenomenon in structured output

Recently, ServiceNow reduces the hallucination phenomenon in structured output through RAG, improves the performance of large language models (LLM), and achieves It achieves out-of-domain generalization while minimizing resource usage.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

The technology involves a RAG system that retrieves relevant JSON objects from an external knowledge base before generating the text. This ensures that the generation process is based on accurate and relevant data.

In addition to RAG, there are five ways to eliminate the illusion of large modelsPicture

By incorporating this pre-retrieval step, the model is less likely to produce false or fabricated information, thereby reducing the phenomenon of hallucinations. Furthermore, this approach allows the use of smaller models without sacrificing performance, making it both efficient and effective.

To learn more about AIGC, please visit:

51CTO AI.x Community

https:// www.51cto.com/aigc/

The above is the detailed content of In addition to RAG, there are five ways to eliminate the illusion of large models. For more information, please follow other related articles on the PHP Chinese website!

Statement
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
AI Therapists Are Here: 14 Groundbreaking Mental Health Tools You Need To KnowAI Therapists Are Here: 14 Groundbreaking Mental Health Tools You Need To KnowApr 30, 2025 am 11:17 AM

While it can’t provide the human connection and intuition of a trained therapist, research has shown that many people are comfortable sharing their worries and concerns with relatively faceless and anonymous AI bots. Whether this is always a good i

Calling AI To The Grocery AisleCalling AI To The Grocery AisleApr 30, 2025 am 11:16 AM

Artificial intelligence (AI), a technology decades in the making, is revolutionizing the food retail industry. From large-scale efficiency gains and cost reductions to streamlined processes across various business functions, AI's impact is undeniabl

Getting Pep Talks From Generative AI To Lift Your SpiritGetting Pep Talks From Generative AI To Lift Your SpiritApr 30, 2025 am 11:15 AM

Let’s talk about it. This analysis of an innovative AI breakthrough is part of my ongoing Forbes column coverage on the latest in AI including identifying and explaining various impactful AI complexities (see the link here). In addition, for my comp

Why AI-Powered Hyper-Personalization Is A Must For All BusinessesWhy AI-Powered Hyper-Personalization Is A Must For All BusinessesApr 30, 2025 am 11:14 AM

Maintaining a professional image requires occasional wardrobe updates. While online shopping is convenient, it lacks the certainty of in-person try-ons. My solution? AI-powered personalization. I envision an AI assistant curating clothing selecti

Forget Duolingo: Google Translate's New AI Feature Teaches LanguagesForget Duolingo: Google Translate's New AI Feature Teaches LanguagesApr 30, 2025 am 11:13 AM

Google Translate adds language learning function According to Android Authority, app expert AssembleDebug has found that the latest version of the Google Translate app contains a new "practice" mode of testing code designed to help users improve their language skills through personalized activities. This feature is currently invisible to users, but AssembleDebug is able to partially activate it and view some of its new user interface elements. When activated, the feature adds a new Graduation Cap icon at the bottom of the screen marked with a "Beta" badge indicating that the "Practice" feature will be released initially in experimental form. The related pop-up prompt shows "Practice the activities tailored for you!", which means Google will generate customized

They're Making TCP/IP For AI, And It's Called NANDAThey're Making TCP/IP For AI, And It's Called NANDAApr 30, 2025 am 11:12 AM

MIT researchers are developing NANDA, a groundbreaking web protocol designed for AI agents. Short for Networked Agents and Decentralized AI, NANDA builds upon Anthropic's Model Context Protocol (MCP) by adding internet capabilities, enabling AI agen

The Prompt: Deepfake Detection Is A Booming BusinessThe Prompt: Deepfake Detection Is A Booming BusinessApr 30, 2025 am 11:11 AM

Meta's Latest Venture: An AI App to Rival ChatGPT Meta, the parent company of Facebook, Instagram, WhatsApp, and Threads, is launching a new AI-powered application. This standalone app, Meta AI, aims to compete directly with OpenAI's ChatGPT. Lever

The Next Two Years In AI Cybersecurity For Business LeadersThe Next Two Years In AI Cybersecurity For Business LeadersApr 30, 2025 am 11:10 AM

Navigating the Rising Tide of AI Cyber Attacks Recently, Jason Clinton, CISO for Anthropic, underscored the emerging risks tied to non-human identities—as machine-to-machine communication proliferates, safeguarding these "identities" become

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

ZendStudio 13.5.1 Mac

ZendStudio 13.5.1 Mac

Powerful PHP integrated development environment

mPDF

mPDF

mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

SecLists

SecLists

SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.

SAP NetWeaver Server Adapter for Eclipse

SAP NetWeaver Server Adapter for Eclipse

Integrate Eclipse with SAP NetWeaver application server.

MinGW - Minimalist GNU for Windows

MinGW - Minimalist GNU for Windows

This project is in the process of being migrated to osdn.net/projects/mingw, you can continue to follow us there. MinGW: A native Windows port of the GNU Compiler Collection (GCC), freely distributable import libraries and header files for building native Windows applications; includes extensions to the MSVC runtime to support C99 functionality. All MinGW software can run on 64-bit Windows platforms.