AgentSkillsCN

research-sop

运用 Claude Code + Perplexity 开展研究方法论。当您需要进行竞争性研究、市场分析,或开展任何需要验证的外部研究时,可调用此技能。核心原则:未知胜过胡说八道。

SKILL.md
--- frontmatter
name: research-sop
description: Research methodology using Claude Code + Perplexity. Use when conducting competitive research, market analysis, or any external research that requires verification. Core principle - Unknown is better than bullshit.
triggers:
  - "research alternatives"
  - "compare * vs *"
  - "evaluate competitors"
  - "find options for"
  - "what are the alternatives"
  - "market research"
  - "competitive analysis"
  - "deep research"
  - "Perplexity"

Research SOP Skill

Standard Operating Procedure for conducting research using Claude Code and Perplexity.

Core Principle: "Unknown is Better Than Bullshit"

NEVER fill in data you haven't verified. Mark unverified fields as "Unknown" rather than guessing or assuming.

  • ❌ Bad: "Audio Transcription: No" (assumed because research didn't mention it)
  • ✅ Good: "Audio Transcription: Unknown" (honest - we didn't verify)

Reasoning: False data causes bad decisions. "Unknown" signals a gap that can be filled later with actual research.


Tool Comparison: Claude Code vs Perplexity

Claude Code - Best For:

StrengthUse Case
Codebase analysisExtracting current technical capabilities from YOUR repos
Context retentionMulti-step tasks where context builds over time
File operationsReading, writing, organizing research outputs
Data processingParsing research results, populating spreadsheets
Tool orchestrationCombining multiple tools (rclone, bash, APIs)
VerificationCross-checking claims against actual code/docs

Limitations:

  • Knowledge cutoff (can't access real-time web data reliably)
  • WebFetch/WebSearch are limited compared to dedicated search
  • Can hallucinate if asked to "research" without actual sources

Perplexity - Best For:

StrengthUse Case
Deep web researchFinding current info about competitors, markets, trends
Citation trackingProvides sources for claims (verify them!)
Real-time dataPricing, funding rounds, recent announcements
Broad exploration"Who else does X?" type queries
Multi-source synthesisCombining info from many websites

Limitations:

  • No access to YOUR codebase or private data
  • Can still hallucinate or misinterpret sources
  • Citations need verification - not always accurate
  • Lacks context about your specific product/situation

Research Workflow SOP

Phase 1: Define What You Know (Claude Code)

Before external research, scan your own repos to establish ground truth:

code
Scan [your repos] to extract current technical state:
- What's implemented vs planned
- Actual capabilities (not marketing claims)
- Architecture and technical decisions

This prevents drift from reality and gives Perplexity accurate context.

Phase 2: External Research (Perplexity)

Craft prompts that:

  1. Provide your context - What you're building, your capabilities
  2. Ask specific questions - Not "tell me about X" but "does X have feature Y?"
  3. Demand sources - "Cite documentation or say 'Not documented'"
  4. List exclusions - What's NOT relevant to avoid noise

Prompt template:

code
Research [topic] for [your context].

## My Current State
[Paste output from Phase 1]

## Specific Questions
For each [entity], verify:
1. [Specific capability] - cite source
2. [Specific capability] - cite source

## Rules
- If not explicitly documented, say "Not documented"
- Do not assume or infer
- Provide URLs for all claims

Phase 3: Verify & Populate (Claude Code)

Take Perplexity output and:

  1. Cross-check claims against cited sources if critical
  2. Mark unverified as "Unknown" - don't propagate assumptions
  3. Populate structured storage (spreadsheet, database, etc.)
  4. Note sources for future verification

Red Flags: When Research is Going Wrong

Red FlagProblemFix
All cells filled, no "Unknown"Likely assumptionsAudit each cell for evidence
Same value for all rowsCopy-paste without verificationCheck each entity individually
No sources citedClaims are unverifiedAsk for citations or mark Unknown
Research contradicts your codeEither research is wrong or code changedVerify against actual implementation
"Seems like" / "Probably"Speculation not factMark as Unknown

Data Quality Tiers

When populating research data, categorize confidence:

TierMeaningAction
VerifiedConfirmed from official docs/codeUse as fact
ReportedFrom credible source (press, reviews)Note source, use cautiously
InferredReasonable assumption, no direct evidenceMark as "Unknown" or note assumption
UnknownNo evidence either wayLeave blank or mark "Unknown"

Reusable Research Artifacts

After research, save:

  1. Raw Perplexity outputs - For future reference
  2. Structured data - Spreadsheet/database with sources
  3. Research prompts - So you can re-run with fresh data
  4. Assumptions log - What you assumed and why (to verify later)

Location convention:

  • Project-specific: [project]/docs/research/
  • Competitor research prompts: [project]/docs/competitor-research-prompt.md

Quick Reference

Use Claude Code when:

  • Analyzing your own code/repos
  • Processing and organizing research outputs
  • Need multi-step workflows with context
  • Populating spreadsheets/databases

Use Perplexity when:

  • Researching external entities (competitors, markets)
  • Need current/real-time information
  • Want cited sources
  • Broad exploration of a topic

Always:

  • Mark unverified data as "Unknown"
  • Note sources for claims
  • Re-verify before major decisions
  • Update research periodically (quarterly minimum for competitors)