creative-thinking-for-research

安装量: 63
排名: #11990

安装

npx skills add https://github.com/orchestra-research/ai-research-skills --skill creative-thinking-for-research
Creative Thinking for Research
Eight empirically grounded frameworks from cognitive science, applied to computer science and AI research. Unlike ad-hoc brainstorming, each framework here is backed by decades of creativity research — from Koestler's bisociation to Kauffman's adjacent possible. They target distinct cognitive operations: combining, reformulating, analogizing, constraining, inverting, abstracting, exploring boundaries, and holding contradictions.
When to Use This Skill
Generating genuinely novel ideas, not incremental extensions of prior work
Feeling trapped in a local optimum of thinking within a single subfield
Wanting to systematically apply creativity heuristics rather than waiting for inspiration
Preparing for a research retreat or PhD-level ideation session
Bridging between fields and seeking structural (not superficial) connections
Do NOT use this skill when
:
You need structured project-level brainstorming workflows (use
brainstorming-research-ideas
)
You have a well-defined problem and need execution help (use domain-specific skills)
You need a literature survey (use
scientific-skills:literature-review
)
Relationship to Brainstorm skill
The brainstorm skill provides operational workflows (diverge → converge → refine) and practical filters. This skill provides the deeper cognitive engines that power creative leaps. Use them together: creative-thinking to generate raw insight, brainstorm to structure and evaluate it.
Framework 1: Combinatorial Creativity (Bisociation)
Novel ideas arise from combining existing concepts in unexpected ways. Arthur Koestler called this
bisociation
— connecting two previously unrelated frames of reference, as distinct from routine association within a single frame.
Why it works
Meta-research consistently shows that breadth of knowledge is a precursor to creative output. People who read across disciplines produce more novel work. The combination itself is the creative act.
In CS Research
:
Biological evolution → optimization (genetic algorithms)
Game theory → networking (mechanism design for routing)
Statistical physics → machine learning (Boltzmann machines, energy-based models)
Linguistics → programming (type theory, formal grammars)
Systematic Bisociation Workflow
:
Select two domains
you have at least passing familiarity with
List core primitives
in each domain (5-10 fundamental concepts per domain)
Create a cross-product matrix
row = concepts from Domain A, column = concepts from Domain B
For each cell
, ask: "What would it mean to apply A's concept to B's problem?"
Filter
Which combinations produce a non-trivial, testable research question?
Validate structural depth
Is the connection mechanistic or merely metaphorical?
Cross-Product Example
:
Caching
Load Balancing
Fault Tolerance
Natural Selection
Evict least-fit entries
Adaptive allocation via fitness
Population-level redundancy
Immune Memory
Learned threat signatures
Distributed detection
Self/non-self discrimination
Symbiosis
Cooperative prefetching
Mutualistic resource sharing
Co-dependent resilience
Quality Test
A strong bisociation is not a surface metaphor ("the network is like a brain") but a structural mapping where the mechanism transfers ("attention mechanisms implement a form of selective gating analogous to cognitive attention filtering").
Self-Check
:
Is the connection structural (mechanisms map) or merely verbal (labels map)?
Does the combination generate testable predictions?
Would an expert in both fields find the connection non-obvious but sound?
Framework 2: Problem Reformulation (Representational Change)
Gestalt psychologists identified that breakthroughs often come not from solving the problem as stated, but from
re-representing the problem itself
. Kaplan and Simon's work on insight shows that changing the problem space — the constraints, the abstraction level, the formalism — is often where creativity lives.
The Key Shift
From "How do I solve this problem?" to "Am I even thinking about this problem correctly?"
Reformulation Strategies
:
Strategy
Example
Change the objective
"Make the algorithm faster" → "Eliminate the need for this computation"
Change the formalism
Graph problem → linear algebra problem (spectral methods)
Change the granularity
Per-token prediction → per-span prediction
Change the agent
"How should the model learn?" → "How should the data teach?" (curriculum learning)
Change the timescale
Real-time optimization → amortized inference
Invert the direction
Forward simulation → inverse problem (learning from observations)
Workflow
:
State your current problem in one sentence
Identify the
hidden assumptions
in that statement:
What formalism are you using? (Could you use a different one?)
What is the objective? (Is it the right objective?)
What level of granularity? (Could you go coarser or finer?)
Who is the agent? (Could you shift perspective?)
For each assumption,
generate the alternative
"What if [opposite assumption]?"
For each alternative, ask: "Does this reformulation make the problem easier, harder, or different in a useful way?"
A reformulation that makes a hard problem easy is often a publishable insight on its own
Classic CS Examples
:
PageRank
Reformulated "find important web pages" from content analysis to graph eigenvalue problem
Dropout
Reformulated "prevent overfitting" from regularization to approximate ensemble
Attention
Reformulated "handle long sequences" from remembering everything to selectively querying
Framework 3: Analogical Reasoning (Structure-Mapping)
Dedre Gentner's
structure-mapping theory
and Kevin Dunbar's studies of real scientists show that analogy is the core engine of scientific creativity. The critical finding: surface-level analogies are common but weak;
structural or relational analogies
— where the deep causal/relational structure maps across domains — produce the most powerful insights.
Dunbar's Finding
In the most successful labs, analogies from distant domains drove the most important discoveries. Nearby analogies refined ideas; distant analogies generated them.
Levels of Analogical Depth
:
Level
Description
Value
Example
Surface
Things look similar
Low
"A neural network is like a brain"
Relational
Relationships between entities match
Medium
"Attention allocation in models parallels resource allocation in economics"
Structural
Deep causal mechanisms map
High
"Diffusion models reverse a thermodynamic process; the math of non-equilibrium stat-mech directly applies"
Structure-Mapping Workflow
:
Describe your problem
using only relational/causal language (strip domain-specific nouns)
Bad: "We need to improve transformer attention efficiency"
Good: "We have a system that must selectively aggregate information from a large set, where relevance is context-dependent and the cost scales quadratically with set size"
Search for structural matches
What other systems selectively aggregate from large sets?
Database query optimization, visual attention in neuroscience, information retrieval, resource allocation
Pick the most distant match
with genuine structural fidelity
Map the solution mechanism
How does the source domain solve this?
Transfer and adapt
What changes when you bring that mechanism into your domain?
Generate predictions
The analogy should tell you something you didn't already know
Validation Checklist
:
Does the mapping preserve causal/relational structure (not just labels)?
Can I identify at least one prediction the analogy makes in my domain?
Would an expert in the source domain confirm the mechanism is correctly understood?
Is the analogy non-obvious to my target audience?
Framework 4: Constraint Manipulation (Boden's Framework)
Margaret Boden's framework distinguishes three forms of creativity based on how they interact with constraints:
Type
Operation
CS Example
Exploratory
Search within the existing conceptual space
Hyperparameter tuning, architecture search within a fixed paradigm
Combinational
Combine elements from different spaces
Multi-task learning, neuro-symbolic methods
Transformational
Change the rules of the space itself
Dropping the assumption that training requires labels (self-supervised learning)
Transformational creativity is the rarest and highest-impact.
It happens when you change what is even considered a valid solution.
Constraint Analysis Workflow
:
List the constraints
of your current approach (5-10 constraints):
Computational: "Must fit in GPU memory"
Methodological: "Requires labeled data"
Architectural: "Uses fixed-length context"
Evaluative: "Measured by accuracy on benchmark X"
Classify each constraint
:
Hard
Physically or logically necessary (cannot violate)
Soft
Convention or historical accident (can question)
Hidden
Not stated but implicitly assumed (most fertile for innovation)
For each soft/hidden constraint
, ask:
What if we relaxed it? (streaming algorithms from relaxing "fits in memory")
What if we tightened it? (efficiency research from tightening compute budgets)
What if we replaced it with a different constraint entirely?
The most productive move
is often exposing and dropping a hidden constraint
Classic Examples of Constraint Transformation
:
"Data must fit in memory" → dropped → streaming algorithms, external memory
"Training requires human labels" → dropped → self-supervised learning
"Models must be deterministic" → dropped → variational methods, diffusion
"Inference must happen in one pass" → dropped → iterative refinement, chain-of-thought
Framework 5: Negation and Inversion
Take a core assumption in your field and negate it. This is formalized in De Bono's lateral thinking and the
TRIZ methodology
from engineering.
The Pattern
"What if [widely held assumption] is wrong, unnecessary, or invertible?"
Systematic Negation Workflow
:
List 5-10 core assumptions
in your subfield (the things "everyone knows")
Negate each one
and ask: What system would you build?
Evaluate each negation
:
Incoherent → discard
Already explored → check if conditions have changed (see brainstorm skill, Framework 5)
Unexplored and coherent → potential research direction
Negation Hall of Fame in CS
:
Assumption
Negation
Result
"We need strong consistency"
What if we don't?
Eventual consistency, CRDTs
"We need exact answers"
What if approximate is fine?
Sketches, LSH, approximate nearest neighbors
"Labels are necessary"
What if we learn without them?
Self-supervised learning, contrastive methods
"More parameters = more compute"
What if we don't use all parameters?
Mixture of Experts, sparse models
"Training and inference are separate"
What if the model keeps learning?
Online learning, test-time training
"Errors must be prevented"
What if we embrace and correct them?
Speculative decoding, self-correction
TRIZ-Inspired Principles for CS
:
TRIZ Principle
CS Application
Inversion
Reverse the process (generative vs. discriminative)
Segmentation
Break monolithic into modular (microservices, mixture of experts)
Merging
Combine separate steps (end-to-end learning)
Universality
One component serves multiple functions (multi-task models)
Nesting
Place one system inside another (meta-learning)
Dynamization
Make static things adaptive (dynamic architectures, adaptive computation)
Framework 6: Abstraction and Generalization Laddering
Moving up and down the abstraction ladder is a fundamental creative act. Polya's heuristics formalize this:
"Can you solve a more general problem? A more specific one? An analogous one?"
Three Moves
:
Move
Question
Outcome
Generalize
"Is my solution a special case of something broader?"
Framework papers, unifying theories
Specialize
"What happens when I add extreme constraints?"
Niche applications, surprising edge cases
Analogize
"Where else does this abstract pattern appear?"
Cross-domain transfer (see Framework 3)
Generalization Workflow
:
State your specific result
Replace each specific element with a variable: "ResNet works for ImageNet" → "Architecture X works for distribution Y"
Ask: Under what conditions does this hold? What is the general principle?
If the general principle is novel → that is the contribution
Specialization Workflow
:
Take a general method
Add extreme constraints: tiny data, huge dimensionality, adversarial inputs, real-time requirements
Ask: Does the method still work? If not, why not?
The failure case often reveals the method's true assumptions
When to Generalize vs. Specialize
:
Generalize when you have results but no explanation
Specialize when you have theory but no grounding
Analogize when you are stuck in either direction
Framework 7: The Adjacent Possible (Kauffman / Johnson)
Stuart Kauffman's concept, popularized by Steven Johnson: innovation happens at the boundary of what is currently reachable — the
adjacent possible
. New ideas become thinkable once their prerequisites exist. This explains why simultaneous independent discovery is so common — multiple people reach the same boundary.
Practical Implication
Map what has recently become possible and explore the space those enablers open.
Adjacent Possible Mapping Workflow
:
List recent enablers
(last 1-3 years):
New hardware capabilities (longer context, faster inference, new accelerators)
New datasets or benchmarks
New open-source tools or frameworks
New theoretical results
New regulatory or social conditions
For each enabler, ask
"What was previously impossible or impractical that this now permits?"
Combine enablers
The most powerful adjacent possibles arise from the intersection of multiple new enablers
Check for competition
If many people can see the same adjacent possible, speed or a unique angle matters
Current Adjacent Possibles (2025-2026)
:
Enabler
Newly Possible
1M+ token context windows
Full-codebase reasoning, book-length analysis
Inference cost drops (100x in 2 years)
Real-time agentic loops, always-on AI assistants
Open-weight models at GPT-4 level
Reproducible research on frontier capabilities
Multimodal models (vision + language + audio)
Unified perception-reasoning systems
Synthetic data at scale
Training data for domains with no natural data
Tool-using models
Research automation, self-improving systems
Timing Signal
If your idea requires technology that doesn't exist yet, it's beyond the adjacent possible — park it. If your idea could have been done 5 years ago, someone probably did — check the literature. The sweet spot is ideas that became feasible in the last 6-18 months.
Framework 8: Janusian and Dialectical Thinking
Albert Rothenberg's studies of eminent creators found that
holding two contradictory ideas simultaneously
is a hallmark of creative thinking. Named after Janus, the two-faced Roman god, this mode of thinking doesn't resolve contradictions by choosing a side — it generates new frameworks that transcend the opposition.
In CS
The most influential results often emerge from tensions previously thought irreconcilable.
Contradiction
Resolution
Impact
Consistency AND Availability (distributed systems)
CAP theorem: formalized the trade-off, then Raft/CRDTs found practical middle grounds
Foundation of distributed systems theory
Security AND Usability
Zero-knowledge proofs: prove knowledge without revealing it
Enabled private computation
Expressiveness AND Tractability
Probabilistic programming: express complex models, automate inference
New programming paradigm
Memorization AND Generalization
Grokking: models memorize first, then generalize with more training
New understanding of learning dynamics
Compression AND Quality
Neural codecs that compress beyond information-theoretic limits via learned priors
Redefined compression research
Dialectical Thinking Workflow
:
Identify a binary
in your field: A vs. B (two approaches, goals, or paradigms treated as opposites)
Resist choosing a side
. Instead ask:
"What would a system look like that achieves both A and B?"
"Under what conditions is the A-B trade-off not fundamental?"
"Is the opposition an artifact of how we formalized the problem?"
Seek synthesis
The resolution often requires a new abstraction that reframes the relationship
Test the synthesis
Can you demonstrate empirically that both goals are achievable?
Self-Check
:
Am I holding the contradiction genuinely (not prematurely resolving it)?
Is the synthesis a new idea, not just a compromise (splitting the difference)?
Does the resolution change how people think about the problem, not just the solution?
Combining Frameworks: A Creative Thinking Protocol
These frameworks are most powerful in combination. Here is a systematic protocol for a deep creative thinking session:
Phase 1: Map the Space (15 min)
Constraint Manipulation
(F4): List all constraints of the current paradigm. Mark which are hard, soft, hidden.
Adjacent Possible
(F7): List recent enablers that change the feasibility landscape.
Phase 2: Generate Disruptions (30 min)
Negation
(F5): Negate 3 soft/hidden constraints. What systems emerge?
Bisociation
(F1): Pick a distant field and create a cross-product matrix with your domain.
Problem Reformulation
(F2): Restate your problem 3 different ways (change objective, formalism, agent).
Phase 3: Deepen Promising Leads (30 min)
Analogical Reasoning
(F3): For each promising idea, find a structural analogy and extract predictions.
Abstraction Laddering
(F6): Move each idea up (generalize) and down (specialize).
Janusian Thinking
(F8): Identify any tensions. Can you synthesize rather than choose?
Phase 4: Evaluate (15 min)
Apply the two-sentence test (from the brainstorm skill):
"
[Domain] currently struggles with [problem] because [reason].
We [approach] by [mechanism], which works because [insight]."
Any idea that survives all four phases and passes the two-sentence test is worth pursuing.
Common Creative Blocks and Unblocking Strategies
Block
Symptom
Framework to Apply
Fixation
Cannot stop thinking about the problem one way
Problem Reformulation (F2) — force a different representation
Tunnel vision
All ideas come from the same subfield
Bisociation (F1) or Analogical Reasoning (F3) — import from elsewhere
Self-censoring
Dismissing ideas as "too weird" before exploring
Negation (F5) — weird is the point; evaluate after generating
Incrementalism
Every idea is "+2% on benchmark X"
Constraint Manipulation (F4) — change the rules, not the parameters
Analysis paralysis
Too many options, cannot commit
Adjacent Possible (F7) — what is feasible right now?
False dichotomy
Stuck choosing between two approaches
Janusian Thinking (F8) — seek synthesis, not selection
Usage Instructions for Agents
When a researcher asks for help with creative thinking or novel ideation:
Assess the block
What kind of thinking are they stuck in? (See Common Creative Blocks table)
Select 2-3 frameworks
based on the block type
Walk through each framework interactively
, asking the researcher to supply domain-specific content
Push for structural depth
If an analogy or combination is surface-level, probe deeper Maintain a running list of all generated ideas, even unusual ones Apply the two-sentence test to candidates that survive exploration Hand off to the brainstorm skill for systematic evaluation (diverge → converge → refine) Key Principles : Generative mode first, evaluative mode second — do not filter prematurely Distant analogies are more valuable than nearby ones, but require more validation The researcher's domain expertise is essential — the agent provides the cognitive scaffolding, not the domain knowledge Encourage the researcher to sit with contradictions rather than resolve them quickly
返回排行榜