Asking LLMs to Verify First is Almost Free Lunch

Asking LLMs to Verify First is Almost Free Lunch

Asking LLMs to Verify First is Almost Free Lunch

Shiguang Wu, Quanming Yao

Department of Electronic Engineering, Tsinghua University

description Abstract

To enhance the reasoning capabilities of Large Language Models (LLMs) without high costs of training, nor extensive test-time sampling, we introduce Verification-First (VF), a strategy that prompts models to verify a provided candidate answer—even a trivial or random one—before generating a solution. This approach triggers a "reverse reasoning" process that is cognitively easier and complementary to standard forward Chain-of-Thought (CoT), effectively invoking the model's critical thinking to reduce logical errors.

psychology Theoretical Foundations

Logical Insight

Verifying an answer is easier than generating a correct answer, providing complementary information to standard CoT

Psychological Insight

Asking one to criticize an answer from others can invoke critical thinking by overcoming egocentrism

Brain and circuit combination showing chain-of-thought reasoning

lightbulb Methodology

Standard CoT Prompting
"Q. [Problem Statement]
A. Let's think step by step..."
Verification-First Prompting
"Q. [Problem Statement]
A possible answer of Q is A'. First verify if A' is correct, then think step by step to find the answer."
Chain-of-thought process visualization

Iter-VF Process: Iterative application of VF using the model's previous answer, creating a Markovian refinement loop that avoids context length and error propagation issues.

analytics Experimental Results

Standard CoT: 70% Accuracy
Verification-First: 85% Accuracy

Performance comparison across various benchmarks

Method Output Tokens (Relative to CoT) Performance Gain
Standard CoT 100% Baseline
Verification-First 120-150% +15% to +25%
Self-Consistency (N=5) 500% +10% to +20%

stars Key Findings

  • VF with random answers consistently outperforms standard CoT with minimal computational overhead
  • Iter-VF outperforms existing TTS strategies under limited computational budgets
  • VF is effective even with thought-hidden commercial LLM services
  • Verification process is the key driver of improvement, not the quality of the candidate answer
Verification-First represents a simple, universal, and powerful method for enhancing LLM reasoning capabilities with minimal additional cost - a "free lunch" in terms of cost versus benefit.

发表评论

人生梦想 - 关注前沿的计算机技术 acejoy.com 🐾 步子哥の博客 🐾 背多分论坛 🐾 知差(chai)网 🐾 DeepracticeX 社区 🐾 老薛主机 🐾 智柴论坛 🐾