--:--
CATEGORIES
AUTHORS

Prompt Repetition Hacks AI: How Saying It Twice Makes LLMs Smarter

Repeating prompts can significantly enhance LLM accuracy. Google's research shows this simple trick outperforms complex AI optimization methods.

Prompt repetition technique showing improved LLM accuracy

A simple copy-paste trick outperforms years of AI optimization rituals—Google just proved that repeating your question makes large language models significantly smarter. The research team’s Prompt Repetition Improves Non-Reasoning LLMs paper revealed that repeating input queries boosts accuracy across Gemini, GPT-4o, Claude, and DeepSeek. In the NameIndex benchmark, Gemini 2.0 Flash-Lite’s accuracy jumped from 21.33% to 97.33% with repetition.

Technically, this works by exploiting the unidirectional 'causal' Transformer architecture. The second prompt iteration gains bidirectional context, creating a 47/70 win rate across non-reasoning benchmarks with 0 losses.

Latency remains minimal due to parallelizable 'prefill' stages, though the technique is ineffective for reasoning tasks like Chain of Thought.

For mid-sized SaaS teams, this means cheaper model scaling options—repeating prompts costs less than upgrading to larger models. However, security risks emerge: repeated injection attacks could exploit this behavior.

Google’s methodology section suggests 'Says it again' is now the new default behavior for inference engines, fundamentally altering how LLMs process input.