EchoPrompt: Instructing the Model to Rephrase Queries for Improved In-context Learning
Just like humans, LLMs understand better when they restate problems in their own words
Just like humans, LLMs understand better when they restate problems in their own words
Simple Query rephrasing in prompts boosts in-context learning across various reasoning tasks for both zero-shot and few-shot.
Problem ๐:
EchoPrompt enhances in-context learning for large language models by incorporating query rephrasing as a subtask. This technique aims to improve model performance across various reasoning tasks.
Solution in this Paper ๐ ๏ธ:
โข Introduces EchoPrompt, a technique that instructs models to rephrase queries before answering
โข Implements EchoPrompt in zero-shot and few-shot settings with various rephrasing structures
โข Adapts EchoPrompt for both standard and chain-of-thought prompting
โข Evaluates performance across multiple reasoning tasks and model families
Key Insights from this Paper ๐ก:
โข Both original and rephrased queries contribute to performance gains
โข EchoPrompt is effective across different model sizes and prompting strategies
โข The technique is robust to irrelevant information in queries
Results ๐:
โข Improves Zero-shot-CoT performance of code-davinci-002 by 5% in numerical tasks and 13% in reading comprehension
โข Enhances few-shot-CoT accuracy on GSM8K from 75.1% to 82.6% with GPT-3.5-turbo
โข Outperforms least-to-most prompting on most benchmarks
โข Maintains improvements even with irrelevant text in queries
Example of EchoPrompt with Compound Sentences.
Standard Prompting approach showcases exemplars with queries and corresponding answering formats. In contrast, the EchoPrompt incorporates a Query-Rephrase step, where the exemplars showcase a rephrased version of the query along with the answering format.



