AI Newsletter Today

Share this post

Reducing Hallucinations in ChatGPT with Chain-of-Verification (CoVe)

www.ainewsletter.today

Discover more from AI Newsletter Today

Latest AI research simplified • Code Tutorials • Weekly AI Roundup
Continue reading
Sign in
AI Research Simplified

Reducing Hallucinations in ChatGPT with Chain-of-Verification (CoVe)

Using CoVe, Llama outperformed leading models like ChatGPT, InstructGPT, and PerplexityAI.

Arjun
Sep 25, 2023
3
Share this post

Reducing Hallucinations in ChatGPT with Chain-of-Verification (CoVe)

www.ainewsletter.today
Share

TL;DR: Hallucinations in LLMs refer to incorrect yet seemingly plausible outputs. The Chain-of-Verification (CoVe) method seeks to mitigate this by having LLMs draft, verify, and refine responses. Llama 65B, using CoVe, surpassed models like ChatGPT in long-form tasks. CoVe's efficacy was notably tested on Wikidata and other tasks, showing an improvement in precision, F1 score, and fact score. Despite its potential, CoVe still has some limitations in completely eradicating hallucinations.

Business Implications

  1. CoVe's improved LLM accuracy can lead to better AI-driven decision-making for businesses.

  2. Enhanced trust in AI outputs can foster stronger customer relationships and brand reputation.

  3. Llama 65B with CoVe might provide superior AI performance and security, differentiating businesses in the market.

  4. Despite CoVe's advancements, businesses should maintain a hybrid approach, combining AI insights with human judgment.

  5. CoVe's proficiency in short-form tasks indicates its potential for efficient chatbots and automated customer interactions.

Here is a sample ChatGPT conversation to show the effectiveness of Chain-of-Verification (CoVe): Chat Link

Subscribe to read more such articles


Hallucinations in LLMs refer to the generation of plausible yet factually incorrect information. As LLMs are trained on an enormous text corpus, spanning billions of tokens, their performance generally improves with an increase in model parameters. Yet, even the most advanced models can falter, especially on tasks less represented in their training data. These errors often appear credible yet are factually incorrect.

Chain-of-Verification (CoVe): A Solution to Hallucinations

Enter the Chain-of-Verification (CoVe) method, a novel approach designed to curb the hallucination issue in LLMs. The CoVe method is a systematic process where the LLM:

  • Draft an initial response

  • Plan verification questions to fact-check the draft

  • Answer the planned verification questions independently to avoid bias

  • Generates a final, verified response

Deep Dive: How CoVe Works

  • Baseline Response: A simple output from LLM is obtained as the starting point, which is typically prone to hallucinations.

  • Plan Verifications: Using the baseline response, LLM generates a series of verification questions that test the factual claims of the baseline responses.

  • Execute Verifications: LLM answers the planned verification questions through several variations like Joint, 2-Step, Factored, and Factor+Revise, each with its unique approach and level of sophistication.

  • Generate Final Verified Response: The improved response that takes verification into account is generated, incorporating any discovered inconsistencies.

CoVe in Action: Experimental Results

CoVe's efficacy was tested across various tasks, including Wikidata, Wikipedia Category List, MultiSpanQA, and long-form biographies. The results are promising:

  • Significant precision improvements in list-based tasks.

  • Enhanced performance in closed book QA, with a 23% F1 score improvement (This represents an improvement to both precision and recall).

  • A 28% increase in fact score for long-form generations.

  • Notably, with CoVe, Llama 65B outperformed leading models like ChatGPT, InstructGPT, and PerplexityAI in long-form generation tasks, marking a significant achievement in the realm of open-source LLMs.

Paper Link.

Check this ChatGPT conversation for prompt examples.

Additional Insights from the Study

The experiments also revealed that short-form verification questions were more accurately answered than long-form ones. Additionally, LLM-based verification questions surpassed heuristic-based ones, and open-ended questions proved more effective than yes/no formats.

Limitations of the CoVe Method

Despite its groundbreaking approach, CoVe isn't without limitations. While it significantly reduces hallucinations, it doesn't eradicate them entirely. There's still a possibility of the model generating misleading information. Moreover, hallucinations might manifest in other forms, such as during incorrect reasoning or when expressing opinions in long-form answers.

Conclusion

The Chain-of-Verification (CoVe) method represents a significant stride in reducing hallucinations in Large Language Models, enhancing their reliability and accuracy across various tasks. By enabling models to verify their responses, CoVe brings us closer to more dependable and error-free artificial intelligence, although some limitations and challenges still need to be addressed.

Thanks for reading the AI Newsletter Today! Subscribe for free to receive new posts and support my work.

3
Share this post

Reducing Hallucinations in ChatGPT with Chain-of-Verification (CoVe)

www.ainewsletter.today
Share
Previous
Next
Comments
Top
New

No posts

Ready for more?

© 2023 AI Newsletter Today
Privacy ∙ Terms ∙ Collection notice
Start WritingGet the app
Substack is the home for great writing