1/16/2026
Share this post:
Export:
The Problem: When you ask an LLM for "an idea," you get the most typical/expected answer. This is called mode collapse.
The Solution: Ask for 5-10 ideas. Then ask for 5-10 MORE ideas that are NOT redundant with the first set. Keep going until the new ideas start overlapping with old ones.
The Insight: When overlap increases, you've "cleared the cognitive market" - exhausted the useful idea space. Now you can stop exploring and start building.
In early 2022, while working on enterprise ideation tasks, I observed a frustrating pattern: LLMs consistently produced stereotypical responses when prompted for single instances. Ask for "a marketing idea" and you get the most generic, expected answer. Ask 10 different times and you get nearly identical responses.
Through experimentation, I developed Clearing the Cognitive Market (CCM) - a technique that systematically explores the full space of possible ideas rather than just the most probable ones.
The core innovation wasn't just asking for lists. It was the overlap analysis stopping criterion: if the second set of ideas has high overlap with the first, you've cleared the market - exhausted the useful idea space. If the sets have low overlap, there's more space to explore.
Large language models are trained to produce the most likely next token. When you ask for "an idea," you get the statistically most common idea - which is, by definition, the most boring one.
This is called mode collapse - the model collapses to a single mode (peak) of the probability distribution instead of sampling from the full range of possibilities.
Think of it this way:
Probability Distribution of "Marketing Ideas"
▲
│ ████
│ ██████
│ ████████ ██
│ ██████████ ████ ██
│████████████ ██████ ████ ██
└─────────────────────────────────→
"Social media" "PR" "Events" (etc.)
↑
LLM always picks this
The LLM keeps returning to "social media campaign" because that's the most probable answer. But the interesting ideas - the ones that could actually differentiate your business - are in the long tail.
Prompt: "Enumerate 10 ideas about [topic]"
Output: Set A = {idea₁, idea₂, ..., idea₁₀}
Human Action: Review and internalize Set A
Prompt: "Enumerate 10 MOAR ideas about [topic] -
DO NOT repeat or be redundant with previous ideas"
Output: Set B = {idea₁₁, idea₁₂, ..., idea₂₀}
Human Action: Calculate overlap |A ∩ B|
IF overlap is HIGH (>30-40%):
→ Market CLEARED - stop exploration
ELSE:
→ More space to explore - continue to Phase 2 with Set C
Let's say you're brainstorming features for a project management app.
"Give me 10 feature ideas for a project management application"
Claude/GPT returns:
These are all... fine. Expected. The kinds of features every competitor already has.
"Give me 10 MORE feature ideas for this project management app - but DO NOT repeat or be redundant with the previous list. Focus on unconventional approaches, edge cases, or things competitors overlook."
Now you might get:
Much more interesting! Several of these are genuinely novel.
"5 more ideas, still non-redundant with everything above"
If you get:
High overlap with previous rounds. The market is cleared.
From 25 generated ideas, you found 3-5 genuinely novel features worth pursuing. Without the CCM technique, you would have stopped at "task lists and Kanban boards" - the same features everyone else has.
I think about LLM prompting in three levels:
Prompt: "Tell me an idea about X"
Result: Mode collapse to single stereotypical response
Prompt: "Tell me 5 ideas about X"
Result: Some diversity, but often repetitive themes
Prompt: "Tell me 10 ideas" → review → "Tell me 10 MOAR, non-redundant"
Result: Explores multiple modes across calls with human guidance
The magic happens at Level 3 because you're not just asking for more - you're explicitly telling the model to avoid its previous outputs, forcing it into less-traveled parts of the probability space.
A crucial insight: this technique requires human judgment. You can't fully automate it.
The human provides:
How do you know when to stop?
Round 1 → Round 2:
Set A Set B
┌─────┐ ┌─────┐
│ │ │ │
│ ○○○│───────│●●● │
│ ○○○│ Low │●●● │
│ │overlap│ │
└─────┘ └─────┘
Conclusion: More to explore!
Round 2 → Round 3:
Set B Set C
┌─────┐ ┌─────┐
│ │▓▓▓▓▓▓▓│ │
│ ●●●│▓▓▓▓▓▓▓│◆◆◆ │
│ ●●●│ High │◆◆◆ │
│ │overlap│ │
└─────┘ └─────┘
Conclusion: Market cleared!
In my experience:
Honestly, you don't need to calculate percentages. After a few rounds, you develop intuition:
That's the market clearing.
When you find a promising idea, drill deeper:
"Idea #7 (decision log) is interesting. Give me 10 specific ways to implement a decision log feature - different UI approaches, data models, or integration patterns."
Now you're clearing the market on a specific sub-problem.
If the model keeps returning similar themes, force a perspective shift:
"Give me 10 ideas, but from these perspectives:
"Give me 10 feature ideas that would be terrible for most companies but perfect for a specific niche. What's the niche, and why would this feature kill it for them?"
This often surfaces unexpected gems.
"What features do ALL project management apps have that users actually hate? Give me 10 ideas for removing or replacing common features."
I've used this technique across dozens of real-world applications:
CCM helps you explore the possibility space, but you still need domain expertise to evaluate which ideas are good. The technique generates candidates; you still have to select.
You could build a script that keeps requesting more ideas, but it would miss the point. The value comes from human judgment guiding the exploration.
If you need the single best answer (not multiple options), standard prompting is fine. CCM is for ideation, brainstorming, and exploration - not for factual queries or routine tasks.
Simply asking for "20 ideas instead of 10" doesn't work as well. The model front-loads obvious answers and then pads with variations. By splitting into rounds with explicit anti-redundancy, you force genuine exploration.
Increasing temperature adds randomness but also reduces quality. CCM maintains quality while increasing diversity because you're guiding the exploration, not just adding noise.
Running 10 separate prompts gives you 10 versions of the "most likely" answer. CCM's explicit anti-redundancy constraint forces the model to avoid its defaults.
Try this today:
Pick a topic you need ideas about
Initial prompt:
"Give me 10 ideas about [topic]. Be specific and actionable."
Review the list. What patterns do you notice? What's missing?
Expansion prompt:
"Give me 10 MORE ideas about [topic]. DO NOT repeat or be redundant with the previous list. Explore unconventional angles, edge cases, or counter-intuitive approaches."
Review and compare. How much overlap? Any surprises?
Repeat until the market clears (high overlap, diminishing novelty)
Select the best ideas and drill down on those
CCM isn't just a prompting trick. It's a mindset shift about how to work with AI.
Most people use LLMs as answer machines: ask question, get answer, done.
CCM treats LLMs as exploration partners: ask for options, review together, push for more, identify when you've exhausted the space, then decide.
This is closer to how you'd work with a smart human collaborator. You wouldn't ask them for "the answer." You'd brainstorm together, challenge each other's assumptions, and keep pushing until you felt confident you'd considered the important possibilities.
The market clearing test gives you a principled stopping rule. Without it, you'd either stop too early (missing good ideas) or keep going forever (wasting time on diminishing returns).
Clearing the Cognitive Market (CCM):
Why it works: Forces the model out of mode collapse by explicitly requiring non-redundant outputs across multiple rounds.
Key insight: The human-in-the-loop is essential. Your judgment guides the exploration and recognizes when it's complete.
Try it today. Next time you need ideas, don't accept the first response. Clear the cognitive market.
This technique works with any LLM - Claude, GPT-4, Gemini, Llama, or any of the 90+ models available today. The key is the human-in-the-loop iterative process, not the specific model.
If you're working with AI at scale - especially in team or enterprise settings - Bike4Mind provides a cognitive workbench that makes CCM-style iterative exploration even more powerful:
For command-line workflows, B4M CLI brings these capabilities to your terminal.
This methodology was developed through three years of production deployment at Bike4Mind. It predates recent academic work on "verbalized sampling" and "distribution-level prompting" while arriving at similar conclusions through practical experimentation.
For the full technical paper with experimental results and enterprise use cases, get in touch or email: erik at bike4mind dot com
Claude on Routing: An AI Reflects on the Human Router Hypothesis
A conversation between Erik Bethke and Claude (Opus 4.5) about intelligence, routing, and a roadmap to AGI through games, shower thoughts, and quantum...
OpenAI O1 Review: Fast, Smart, but Surprisingly Reserved
A hands-on review of OpenAI's O1 'unlimited' model, comparing it with Claude and exploring its unique personality quirks.
Building an AI Survey with SST v3, Next.js, and Claude
From concept to production in 37 minutes: Building a full-stack survey application with SST v3, Next.js, and AI pair programming
Get notified when I publish new blog posts about game development, AI, entrepreneurship, and technology. No spam, unsubscribe anytime.