brains-trust

安装量: 99
排名: #8385

安装

npx skills add https://github.com/jezweb/claude-skills --skill brains-trust
Brains Trust
Consult other leading AI models for a second opinion. Not limited to code — works for architecture, strategy, prompting, debugging, writing, or any question where a fresh perspective helps.
Defaults (When User Just Says "Brains Trust")
If the user triggers this skill without specifying what to consult about, apply these defaults:
Pattern
Consensus (2 models from different providers) — it's called "brains trust", not "single opinion"
Scope
Whatever Claude has been working on in the current session. Look at recent context: files edited, decisions made, architecture discussed, problems being solved.
Mode
Infer from context:
Recently wrote/edited code →
Code Review
In a planning or design discussion →
Architecture
Debugging something →
Debug
Building prompts or skills →
Prompting
No clear signal →
General
(ask: "what are we missing? what are our blind spots?")
Models
Pick the newest pro-tier model from 2 different providers (check
models.flared.au
). Prefer diversity: e.g. one Google + one OpenAI, or one Qwen + one Google. Never two from the same provider.
Prompt focus
"Review what we've been working on. What are we missing? What could be improved? What blind spots might we have? Are there simpler approaches we haven't considered?" Trigger → Default Mapping Trigger Default pattern Default scope "brains trust" Consensus (2 models) Current session work "second opinion" Single (1 model) Current session work "ask gemini" / "ask gpt" Single (specified provider) Current session work "peer review" Consensus (2 models) Recently changed files "challenge this" / "devil's advocate" Devil's advocate (1 model) Claude's current position The user can always override by being specific: "brains trust this config file", "ask gemini about the auth approach", etc. Setup Set at least one API key as an environment variable:

Recommended — one key covers all providers

export OPENROUTER_API_KEY = "your-key"

Optional — direct access (often faster/cheaper)

export
GEMINI_API_KEY
=
"your-key"
export
OPENAI_API_KEY
=
"your-key"
OpenRouter is the universal path — one key gives access to Gemini, GPT, Qwen, DeepSeek, Llama, Mistral, and more.
Current Models
Do not use hardcoded model IDs.
Before every consultation, fetch the current leading models:
https://models.flared.au/llms.txt
This is a live-updated, curated list of ~40 leading models from 11 providers, filtered from OpenRouter's full catalogue. Use it to pick the right model for the task.
For programmatic use in the generated Python script:
https://models.flared.au/json
Consultation Patterns
Pattern
Default for
What happens
Consensus
"brains trust", "peer review"
Ask 2 models from different providers in parallel, compare where they agree/disagree
Single
"second opinion", "ask gemini", "ask gpt"
Ask one model, synthesise with your own view
Devil's advocate
"challenge this", "devil's advocate"
Ask a model to explicitly argue against your current position
For consensus, always pick models from different providers (e.g. one Google + one Qwen) for maximum diversity of perspective.
Modes
Mode
When
Model tier
Code Review
Review files for bugs, patterns, security
Flash
Architecture
Design decisions, trade-offs
Pro
Debug
Stuck after 2+ failed attempts
Flash
Security
Vulnerability scan
Pro
Strategy
Business, product, approach decisions
Pro
Prompting
Improve prompts, system prompts, KB files
Flash
General
Any question, brainstorm, challenge
Flash
Pro tier
The most capable model from the chosen provider (e.g.
google/gemini-3.1-pro-preview
,
openai/gpt-5.4
).
Flash tier
Fast, cheaper models for straightforward analysis (e.g. google/gemini-3-flash-preview , qwen/qwen3.5-flash-02-23 ). Workflow Detect available keys — check OPENROUTER_API_KEY , GEMINI_API_KEY , OPENAI_API_KEY in environment. If none found, show setup instructions and stop. Fetch current models — WebFetch https://models.flared.au/llms.txt and pick appropriate models based on mode (pro vs flash) and consultation pattern (single vs consensus). If user requested a specific provider ("ask gemini"), use that. Read target files into context (if code-related). For non-code questions (strategy, prompting, general), skip file reading. Build prompt using the AI-to-AI template from references/prompt-templates.md . Include file contents inline with --- filename --- separators. Do not set output token limits — let models reason fully. Create consultation directory at .jez/artifacts/brains-trust/{timestamp}-{topic}/ (e.g. 2026-03-10-1423-auth-architecture/ ). Write the prompt to prompt.txt inside it — never pass code inline via bash arguments (shell escaping breaks it). Generate and run Python script at .jez/scripts/brains-trust.py using patterns from references/provider-api-patterns.md : Reads prompt from the consultation directory's prompt.txt Calls the selected API(s) For consensus mode: calls multiple APIs in parallel using concurrent.futures Saves each response to {model}.md in the consultation directory Prints results to stdout Synthesise — read the responses, present findings to the user. Note where models agree and disagree. Add your own perspective (agree/disagree with reasoning). Let the user decide what to act on. When to Use Good use cases : Before committing major architectural changes When stuck debugging after multiple attempts Architecture decisions with multiple valid options Reviewing security-sensitive code Challenging your own assumptions on strategy or approach Improving system prompts or KB files Any time you want a fresh perspective Avoid using for : Simple syntax checks (Claude handles these) Every single edit (too slow, costs money) Questions with obvious, well-known answers Critical Rules Never hardcode model IDs — always fetch from models.flared.au first Never cap output tokens — don't set max_tokens or maxOutputTokens Always write prompts to file — never pass via bash arguments Include file contents inline — attach code context directly in the prompt Use AI-to-AI framing — the model is advising Claude, not talking to the human Print progress to stderr — the Python script must print status updates ( Calling gemini-2.5-pro... , Received response from qwen3.5-plus. ) so the user knows it's working during the 30-90 second wait Reference Files When Read Building prompts for any mode references/prompt-templates.md Generating the Python API call script references/provider-api-patterns.md
返回排行榜