


In addition to RAG, there are five ways to eliminate the illusion of large models
Produced by 51CTO Technology Stack (WeChat ID: blog51cto)
It is well known that LLM can produce hallucinations - that is, generate incorrect, misleading or meaningless information.
Interestingly, some people, such as OpenAI CEO Sam Altman, view the imagination of AI as creativity, while others believe that imagination may help make new scientific discoveries.
In most cases, however, it is crucial to provide the correct answer, and hallucinations are not a feature, but a flaw.
So, how to reduce the illusion of LLM? Long context? RAG? Fine-tuning?
In fact, long context LLMs are not foolproof, vector search RAG is not satisfactory, and fine-tuning comes with its own challenges and limitations.
The following are some advanced techniques that can be used to reduce the LLM illusion.
1. Advanced prompts
There is indeed a lot of discussion about whether using better or more advanced prompts can solve the hallucination problem of large language models (LLM).
Picture
Some people think that writing more detailed prompt words will not help solve the (hallucination) problem, but the co-founder of Google Brain (Google Brain) But people like Andrew Ng saw the potential. They proposed a new method that uses deep learning technology to generate prompt words to help people solve problems better. This method utilizes a large amount of data and powerful computing power to automatically generate prompt words related to the problem, thereby improving the efficiency of problem solving. Although the field
Ng believes that the inference capabilities of GPT-4 and other advanced models make them very good at interpreting complex prompt words with detailed instructions.
Picture
“With multi-example learning, developers can give dozens, or even hundreds, of examples in a prompt word, which is better than less Learning by example is more effective,” he writes.
Picture
In order to improve prompt words, many new developments are constantly emerging. For example, Anthropic released a new The “Prompt Generator” tool converts simple descriptions into high-level prompts optimized for large language models (LLMs). Through the Anthropic console, you can generate prompt words for production.
Recently, Marc Andreessen also said that with the right prompts, we can unlock the potential super genius in AI models. "Prompting techniques in different areas could unlock this potential super-genius," he added.
2.Meta AI’s Chain-of-Verification (CoVe)
Meta AI’s Chain-of-Verification (CoVe) is another technology. This approach reduces hallucinations in large language models (LLMs) by breaking down fact-checking into manageable steps, improving response accuracy, and aligning with human-driven fact-checking processes.
Picture
CoVe involves generating an initial response, planning validation questions, answering those questions independently, and generating a final validated response. This approach significantly improves the accuracy of the model by systematically validating and correcting its output.
It improves performance in a variety of tasks such as list-based questions, closed-book question answering, and long-form text generation by reducing hallucinations and increasing factual correctness.
3. Knowledge Graph
RAG (Retrieval Enhanced Generation) is no longer limited to vector database matching. Many advanced RAG technologies have been introduced to significantly improve the retrieval effect.
Picture
For example, integrate knowledge graphs (KGs) into RAG. By leveraging structured and interconnected data in knowledge graphs, the reasoning capabilities of current RAG systems can be greatly enhanced.
4.Raptor
Another technique is Raptor, which handles problems that span multiple documents by creating a higher level of abstraction. It is particularly useful when answering queries involving multiple document concepts.
Picture
Approaches like Raptor fit well with long-context large language models (LLMs) because you can directly embed the entire document without chunking .
This method reduces the hallucination phenomenon by integrating an external retrieval mechanism with the transformer model. When a query is received, Raptor first retrieves relevant and verified information from external knowledge bases.
These retrieved data are then embedded into the context of the model along with the original query. By basing the model's responses on facts and relevant information, Raptor ensures that the content generated is both accurate and contextual.
5. Conformal Abstention
The paper "Relieving the Hallucination Phenomenon of Large Language Models through Conformal Abstention" introduces a method to determine the model by applying conformal prediction technology Methods to reduce hallucinations in large language models (LLMs) when responses should be avoided.
Picture
By using self-consistency to evaluate the similarity of responses and leveraging conformal prediction for strict guarantees, this method ensures that the model only Only respond when you are confident in its accuracy.
This method effectively limits the incidence of hallucinations while maintaining a balanced withdrawal rate, which is especially beneficial for tasks that require long answers. It significantly improves the reliability of model output by avoiding erroneous or illogical responses.
6.RAG reduces the hallucination phenomenon in structured output
Recently, ServiceNow reduces the hallucination phenomenon in structured output through RAG, improves the performance of large language models (LLM), and achieves It achieves out-of-domain generalization while minimizing resource usage.
Picture
The technology involves a RAG system that retrieves relevant JSON objects from an external knowledge base before generating the text. This ensures that the generation process is based on accurate and relevant data.
Picture
By incorporating this pre-retrieval step, the model is less likely to produce false or fabricated information, thereby reducing the phenomenon of hallucinations. Furthermore, this approach allows the use of smaller models without sacrificing performance, making it both efficient and effective.
To learn more about AIGC, please visit:
https:// www.51cto.com/aigc/
The above is the detailed content of In addition to RAG, there are five ways to eliminate the illusion of large models. For more information, please follow other related articles on the PHP Chinese website!

While it can’t provide the human connection and intuition of a trained therapist, research has shown that many people are comfortable sharing their worries and concerns with relatively faceless and anonymous AI bots. Whether this is always a good i

Artificial intelligence (AI), a technology decades in the making, is revolutionizing the food retail industry. From large-scale efficiency gains and cost reductions to streamlined processes across various business functions, AI's impact is undeniabl

Let’s talk about it. This analysis of an innovative AI breakthrough is part of my ongoing Forbes column coverage on the latest in AI including identifying and explaining various impactful AI complexities (see the link here). In addition, for my comp

Maintaining a professional image requires occasional wardrobe updates. While online shopping is convenient, it lacks the certainty of in-person try-ons. My solution? AI-powered personalization. I envision an AI assistant curating clothing selecti

Google Translate adds language learning function According to Android Authority, app expert AssembleDebug has found that the latest version of the Google Translate app contains a new "practice" mode of testing code designed to help users improve their language skills through personalized activities. This feature is currently invisible to users, but AssembleDebug is able to partially activate it and view some of its new user interface elements. When activated, the feature adds a new Graduation Cap icon at the bottom of the screen marked with a "Beta" badge indicating that the "Practice" feature will be released initially in experimental form. The related pop-up prompt shows "Practice the activities tailored for you!", which means Google will generate customized

MIT researchers are developing NANDA, a groundbreaking web protocol designed for AI agents. Short for Networked Agents and Decentralized AI, NANDA builds upon Anthropic's Model Context Protocol (MCP) by adding internet capabilities, enabling AI agen

Meta's Latest Venture: An AI App to Rival ChatGPT Meta, the parent company of Facebook, Instagram, WhatsApp, and Threads, is launching a new AI-powered application. This standalone app, Meta AI, aims to compete directly with OpenAI's ChatGPT. Lever

Navigating the Rising Tide of AI Cyber Attacks Recently, Jason Clinton, CISO for Anthropic, underscored the emerging risks tied to non-human identities—as machine-to-machine communication proliferates, safeguarding these "identities" become


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

ZendStudio 13.5.1 Mac
Powerful PHP integrated development environment

mPDF
mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

SecLists
SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.

SAP NetWeaver Server Adapter for Eclipse
Integrate Eclipse with SAP NetWeaver application server.

MinGW - Minimalist GNU for Windows
This project is in the process of being migrated to osdn.net/projects/mingw, you can continue to follow us there. MinGW: A native Windows port of the GNU Compiler Collection (GCC), freely distributable import libraries and header files for building native Windows applications; includes extensions to the MSVC runtime to support C99 functionality. All MinGW software can run on 64-bit Windows platforms.
