Workflow
提示词重复
icon
Search documents
Gemini准确率从21%飙到97%,谷歌只用了这一招:复制粘贴
3 6 Ke· 2026-01-19 12:33
Core Insights - A simple prompt repetition technique can significantly enhance the accuracy of large language models (LLMs) from 21.33% to 97.33% without requiring reasoning capabilities [1][7][28] - Google Research's findings challenge the complexity of previous prompt engineering methods, suggesting that straightforward approaches can yield substantial improvements in model performance [2][5][48] Group 1: Research Findings - The study demonstrated that repeating the input question can lead to a remarkable accuracy increase, with a maximum improvement of 76 percentage points in non-reasoning tasks [1][7] - In tests across seven common benchmarks and various mainstream models, the prompt repetition method outperformed baseline methods in 47 out of 70 comparisons, with no losses [5][7] - The specific example of Gemini 2.0 Flash-Lite showed an accuracy jump from 21.33% to 97.33% when the input was repeated, illustrating the effectiveness of this technique [7][28] Group 2: Mechanism Behind the Technique - The underlying reason for the effectiveness of prompt repetition is linked to the "Causal Blind Spot" in transformer models, which limits their ability to utilize previous context effectively [9][10] - By repeating the prompt, the model gains a "look-back" opportunity, allowing it to leverage previously processed information, thus enhancing its performance on tasks requiring precise information retrieval [14][15] Group 3: Implications for Developers - The findings suggest that developers do not need to invest in larger, more expensive models to achieve high accuracy; instead, they can optimize smaller models through prompt repetition [26][28] - This technique allows for maintaining efficiency, as the time cost associated with repeating prompts is negligible due to the parallel processing capabilities of modern GPUs [18][24] Group 4: Limitations and Considerations - While prompt repetition is effective for non-reasoning tasks, it is not suitable for tasks requiring step-by-step reasoning, where traditional methods like "Chain of Thought" may still be necessary [31][33] - The research indicates that combining prompt repetition with reasoning techniques may not yield the desired results, as models may already internally repeat prompts during reasoning processes [37][39]
Gemini准确率从21%飙到97%!谷歌只用了这一招:复制粘贴
猿大侠· 2026-01-19 04:11
Core Insights - A recent study by Google Research reveals that simply repeating a question can significantly enhance the accuracy of large language models (LLMs) from 21.33% to 97.33% without requiring reasoning capabilities [1][4][18] - This technique, termed "prompt repetition," challenges the need for complex prompting strategies like "Chain of Thought" and "Multi-shot" [1][9][10] Group 1: Effectiveness of Prompt Repetition - The study demonstrated that prompt repetition outperformed baseline methods in 47 out of 70 tests, with no losses recorded [12][13] - In a specific test involving identifying the 25th name from a list of 50, the accuracy of Gemini 2.0 Flash-Lite improved from 21.33% to 97.33% through repetition [16][18] - The technique provides a "look-back" opportunity for models, allowing them to utilize previously seen information, thus enhancing performance [29][32] Group 2: Efficiency and Cost-Effectiveness - Prompt repetition does not significantly impact generation speed, as the processing of repeated prompts is highly parallelizable [36][40] - This finding suggests that developers can achieve high accuracy without the need for larger, more expensive models, making it a cost-effective solution [41][42] - The ability to enhance smaller models' performance to match or exceed that of larger models represents a significant advancement in AI technology [42] Group 3: Limitations and Safety Considerations - While effective for retrieval tasks, prompt repetition is not suitable for reasoning tasks, where models may already internally repeat the prompt [46][52] - The increased attention mechanism from repetition could potentially amplify certain instructions, raising security concerns regarding model vulnerabilities [56][58] - Developers are encouraged to consider the implications of prompt repetition on both model performance and security, potentially using it as a defensive strategy [60][61]