Files
novelty-seeking/research/theoretical_framework.md
gbanyan 26a56a2a07 feat: Enhance patent search and update research documentation
- Improve patent search service with expanded functionality
- Update PatentSearchPanel UI component
- Add new research_report.md
- Update experimental protocol, literature review, paper outline, and theoretical framework

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
2026-01-19 15:52:33 +08:00

12 KiB
Raw Permalink Blame History

Theoretical Framework: Expert-Augmented LLM Ideation

The Core Problem: LLM "Semantic Gravity"

What is Semantic Gravity?

When LLMs generate creative ideas directly, they exhibit a phenomenon we term "semantic gravity" - the tendency to generate outputs that cluster around high-probability regions of their training distribution.

Direct LLM Generation:
  Input: "Generate creative ideas for a chair"

  LLM Process:
    P(idea | "chair") → samples from training distribution

  Result:
    - "Ergonomic office chair" (high probability)
    - "Foldable portable chair" (high probability)
    - "Eco-friendly bamboo chair" (moderate probability)

  Problem:
    → Ideas cluster in predictable semantic neighborhoods
    → Limited exploration of distant conceptual spaces
    → "Creative" outputs are interpolations, not extrapolations

Why Does This Happen?

  1. Statistical Pattern Learning: LLMs learn co-occurrence patterns from training data
  2. Mode Collapse: When asked to be "creative," LLMs sample from the distribution of "creative ideas" they've seen
  3. Relevance Trap: Strong associations dominate weak ones (chair→furniture >> chair→marine biology)
  4. Prototype Bias: Outputs gravitate toward category prototypes, not edge cases

The Solution: Expert Perspective Transformation

Theoretical Basis

Our approach draws from three key theoretical foundations:

1. Semantic Distance Theory (Mednick, 1962)

"Creative thinking involves connecting weakly related, remote concepts in semantic memory."

Key insight: Creativity correlates with semantic distance. The farther the conceptual "jump," the more creative the result.

Our application: Expert perspectives force semantic jumps that LLMs wouldn't naturally make.

Without Expert:
  "Chair" → furniture, sitting, comfort, design
  Semantic distance: SHORT

With Marine Biologist Expert:
  "Chair" → underwater pressure, coral structure, buoyancy, bioluminescence
  Semantic distance: LONG

Result: Novel ideas like "pressure-adaptive seating" or "coral-inspired structural support"

The Semantic Distance Tradeoff

However, semantic distance is not always beneficial. There exists a tradeoff:

Semantic Distance Spectrum:

Too Close                    Optimal Zone                    Too Far
(Semantic Gravity)           (Creative)                      (Hallucination)
├────────────────────────────┼────────────────────────────────┼────────────────────────────┤
"Ergonomic office chair"     "Pressure-adaptive seating"     "Quantum-entangled
                             "Coral-inspired support"         chair consciousness"

High usefulness              High novelty + useful            High novelty, nonsense
Low novelty                                                   Low usefulness

Our Design Choice: Context-free keyword generation (Stage 1 excludes original query) intentionally pushes toward the "far" end to maximize novelty. Stage 2 re-introduces query context to ground the ideas.

Research Question: What is the hallucination/nonsense rate of this approach, and is the tradeoff worthwhile?

2. Conceptual Blending Theory (Fauconnier & Turner, 2002)

"Creative products emerge from blending elements of two input spaces into a novel integrated space."

The blending process:

  1. Input Space 1: The target concept (e.g., "chair")
  2. Input Space 2: The expert's domain knowledge (e.g., marine biology)
  3. Generic Space: Abstract structure shared by both
  4. Blended Space: Novel integration of elements from both inputs

Our application: Each expert provides a distinct input space for systematic blending.

┌─────────────────┐     ┌─────────────────┐
│   Input 1       │     │   Input 2       │
│   "Chair"       │     │ Marine Biology  │
│   - support     │     │ - pressure      │
│   - sitting     │     │ - buoyancy      │
│   - comfort     │     │ - adaptation    │
└────────┬────────┘     └────────┬────────┘
         │                       │
         └───────────┬───────────┘
                     ▼
         ┌─────────────────────┐
         │   Blended Space     │
         │   Novel Chair Ideas │
         │   - pressure-adapt  │
         │   - buoyant support │
         │   - bio-adaptive    │
         └─────────────────────┘

3. Design Fixation Breaking (Jansson & Smith, 1991)

"Design fixation is blind adherence to initial ideas, limiting creative output."

Fixation occurs because:

  • Knowledge is organized around category prototypes
  • Prototypes require less cognitive effort to access
  • Initial examples anchor subsequent ideation

Our application: Expert perspectives act as "defixation triggers" by activating non-prototype knowledge.

Without Intervention:
  Prototype: "standard four-legged chair"
  Fixation: Variations on four-legged design

With Expert Intervention:
  Archaeologist: "Ancient people sat differently..."
  Dance Therapist: "Seating affects movement expression..."

  Fixation Broken: Entirely new seating paradigms explored

The Multi-Expert Aggregation Model

From "Wisdom of Crowds" to "Inner Crowd"

Research shows that groups generate more diverse ideas because each member brings different perspectives. Our system simulates this "crowd wisdom" through multiple expert personas:

Traditional Crowd:
  Person 1 → Ideas from perspective 1
  Person 2 → Ideas from perspective 2
  Person 3 → Ideas from perspective 3
  Aggregation → Diverse idea pool

Our "Inner Crowd":
  LLM + Expert 1 Persona → Ideas from perspective 1
  LLM + Expert 2 Persona → Ideas from perspective 2
  LLM + Expert 3 Persona → Ideas from perspective 3
  Aggregation → Diverse idea pool (simulated crowd)

Why This Approach Works: Two Complementary Mechanisms

Factor 1: Attribute Decomposition

  • Structures the problem space before creative exploration
  • Prevents premature fixation on holistic solutions
  • Ensures coverage across different aspects of the target concept

Factor 2: Expert Perspectives

  • Different experts activate different semantic regions
  • Forces semantic jumps that LLMs wouldn't naturally make
  • Each expert provides a distinct input space for conceptual blending

Combined Effect (Interaction)

  • Experts are more effective when given structured attributes to transform
  • Attributes without expert perspectives still generate predictable ideas
  • The combination creates systematic exploration of remote conceptual spaces

The Complete Pipeline

Stage 1: Attribute Decomposition

Purpose: Structure the problem space before creative exploration

Input: "Innovative chair design"

Output:
  Categories: [Material, Function, Usage, User Group]

  Material: [wood, metal, fabric, composite]
  Function: [support, comfort, mobility, storage]
  Usage: [office, home, outdoor, medical]
  User Group: [children, elderly, professionals, athletes]

Theoretical basis: Structured decomposition prevents premature fixation on holistic solutions.

Stage 2: Expert Team Generation

Purpose: Assemble diverse perspectives for maximum semantic coverage

Strategies:
  1. LLM-Generated: Query-specific, prioritizes unconventional experts
  2. Curated: Pre-selected high-quality occupations
  3. External Sources: DBpedia, Wikidata for broad coverage

Diversity Optimization:
  - Domain spread (arts, science, trades, services)
  - Expertise level variation
  - Cultural/geographic diversity

Stage 3: Expert Transformation

Purpose: Apply each expert's perspective to each attribute

For each (attribute, expert) pair:

  Input: "Chair comfort" + "Marine Biologist"

  LLM Prompt:
    "As a marine biologist, how might you reimagine
     chair comfort using principles from your field?"

  Output: Keywords + Descriptions
    - "Pressure-distributed seating inspired by deep-sea fish"
    - "Buoyancy-assisted support reducing pressure points"

Stage 4: Deduplication

Purpose: Ensure idea set is truly diverse, not just numerous

Methods:
  1. Embedding-based: Fast cosine similarity clustering
  2. LLM-based: Semantic pairwise comparison (more accurate)

Output:
  - Unique ideas grouped by similarity
  - Representative idea selected from each cluster
  - Diversity metrics computed

Stage 5: Novelty Validation

Purpose: Ground novelty in real-world uniqueness

Process:
  - Search patent databases for similar concepts
  - Compute overlap scores
  - Flag ideas with high existing coverage

Output:
  - Novelty score per idea
  - Patent overlap rate for idea set

Testable Hypotheses (2×2 Factorial Design)

Our experimental design manipulates two independent factors:

  1. Attribute Decomposition: With / Without
  2. Expert Perspectives: With / Without

H1: Main Effect of Attribute Decomposition

Conditions with attribute decomposition produce higher semantic diversity than those without.

Prediction: (Attribute-Only + Full Pipeline) > (Direct + Expert-Only) Measurement: Mean pairwise cosine distance between idea embeddings

H2: Main Effect of Expert Perspectives

Conditions with expert perspectives produce higher semantic diversity than those without.

Prediction: (Expert-Only + Full Pipeline) > (Direct + Attribute-Only) Measurement: Mean pairwise cosine distance between idea embeddings

H3: Interaction Effect

The combination of attributes and experts produces super-additive benefits.

Prediction: Full Pipeline > (Attribute-Only + Expert-Only - Direct) Rationale: Experts are more effective when given structured problem decomposition to work with. Measurement: Interaction term in 2×2 ANOVA

H4: Novelty

The Full Pipeline produces ideas with lowest patent overlap.

Prediction: Full Pipeline has highest novelty rate across all conditions Measurement: Percentage of ideas without existing patent matches

H5: Expert vs Random Control

Expert perspectives outperform random word perspectives.

Prediction: Expert-Only > Random-Perspective Rationale: Validates that domain knowledge (not just any perspective shift) drives improvement Measurement: Semantic diversity and human creativity ratings


Expected Contributions

  1. Theoretical: Formalization of "semantic gravity" as LLM creativity limitation
  2. Methodological: Expert-augmented ideation pipeline with evaluation framework
  3. Empirical: Quantitative evidence for multi-expert creativity enhancement
  4. Practical: Open-source system for innovation ideation

Key Differentiator: Attribute Decomposition

PersonaFlow (2024):        Query → Experts → Ideas
Our Approach:              Query → Attributes → (Attributes × Experts) → Ideas

Why this matters: Attribute decomposition provides scaffolding that makes expert perspectives more effective. An expert seeing "chair materials" generates more focused ideas than an expert seeing just "chair."

Comparison Table

Approach Limitation Our Advantage
Direct LLM generation Semantic gravity, fixation Two-factor enhancement (attributes + experts)
PersonaFlow (2024) No problem structure, experts see whole query Attribute decomposition amplifies expert effect
PopBlends (2023) Two-concept blending only Systematic attribute × expert exploration
BILLY (2025) Cannot isolate what helps 2×2 factorial design isolates contributions
Persona prompting alone Random coverage Systematic coverage via attribute × expert matrix

What We Can Answer That PersonaFlow Cannot

  1. Does problem structure alone help? (Attribute-Only vs Direct)
  2. Do experts help beyond structure? (Full Pipeline vs Attribute-Only)
  3. Is there an interaction effect? (Full Pipeline > Attribute-Only + Expert-Only - Direct)

PersonaFlow showed experts help, but never tested whether structuring the problem first makes experts more effective.