Install
openclaw skills install scientific-inquiryRigorous evidence-based inquiry: decompose fuzzy questions, retrieve & grade evidence (S/A/B/C/D), cross-validate, and output conclusions with confidence intervals. Includes Step 0 user confirmation to prevent direction drift.
openclaw skills install scientific-inquirySecurity Notice: This skill uses self-modification (via
skill_manage) but ONLY when the user explicitly commands it. See the "Controlled Self-Evolution" section for details. This prevents prompt injection and unintended auto-modification.
Activate this skill when the user asks any of the following:
Even simple requests (like "check this stat") activate this skill if they involve systematic information gathering.
Upon receiving a question, do NOT start searching yet. First output a research plan template:
📋 Research Plan
Question: [Restate the original question to confirm alignment]
Research type: Fact-check / Data research / Industry study / Comparison / Trend analysis
Sub-questions:
- [Sub-question A] — Verifiability: High/Medium/Low → [Expected sources]
- [Sub-question B] — Verifiability: High/Medium/Low → [Expected sources]
- [Sub-question C] — Verifiability: High/Medium/Low → [Expected sources]
Methodology:
- Primary search path: [Specific tools/APIs/databases]
- Keywords: [Search terms]
- Fallback if key data is unavailable: [Alternative approach]
Expected output:
- Expected confidence: High/Medium/Low
- Main uncertainties: [Anticipated blind spots]
✅ Does this direction look good? Let me know and I'll proceed with Step 1-4.
Do NOT make any retrieval tool calls until the user confirms.
Break the fuzzy question into verifiable atomic statements. For each:
Before formal research, check these prerequisites:
Classic failure mode: User asks "will Huawei phones get more expensive?" You analyze storage cost trends for 30 minutes. Meanwhile, the Pura 90 already launched with published pricing. You're predicting history.
Before committing to a search tool, quickly test availability:
curl -sL to Google/Bing/DuckDuckGo; distinguish CAPTCHA from timeout(empty page) or ERR_TIMED_OUT) → network/proxy issue
curl to a simple HTTP target(empty page) sometimes resolves after pressing Enter/submitting the search formThis step prevents wasted calls on dead search channels. If all search engines are blocked, video platform titles + vertical media browsing is 10x more productive than retrying Google.
Every piece of evidence MUST be annotated with source and grade. See the "Evidence Classification Discipline" section for detailed definitions.
Prioritize S/A-grade evidence; B/C are supplementary only.
S-grade: Primary academic literature / Official statistics / Raw data APIs
A-grade: Authoritative media / Professional reports / Fully cited secondary sources
B-grade: Industry analysis / Forum discussions / Indirect data
C-grade: Social media / Single samples / Non-professional interpretations
D-grade: No source / Rumors / Obvious conflicts of interest
Present findings as an evidence table:
| Evidence | Source | URL | Grade | Sub-question |
|---|---|---|---|---|
| ... | ... | ... | ... | ... |
Source URLs are mandatory. A bare site name (e.g., "YouTube") is not a valid source. Even search engine results should link to the search page or specific result.
When mainstream search engines are blocked or return empty results:
1️⃣ Video platform search — YouTube (for pricing/product info), or local equivalents
2️⃣ Direct access to vertical media
3️⃣ E-commerce platforms
4️⃣ Social media
5️⃣ Text-mode search engines
Priority: Video platform titles > Vertical media > E-commerce > Social media. Video title info density and timeliness often exceed other sources for consumer products.
For each sub-question:
Two-block output:
Block A — Claim Verification Report (one line per key finding)
✅ CONFIRMED: 「Pura 90 starts at ¥4,699」→ 5 creator video titles agree + financial media report
⚠️ UNVERIFIABLE: 「Huawei stockpiled 100M NAND chips」→ single comment section post (D-grade), no media confirmation
❌ CONTRADICTED: 「Pura 90 will be more expensive than Pura 80」→ actual launch price ¥4,699, same as predecessor
Block B — Overall Judgment
Proposition: [One-sentence restatement]
Confidence:
✅ High (≥80%) — Multiple S/A-grade evidence consistent
⚠️ Medium (50-80%) — Key data gaps exist
❌ Low (<50%) — Mostly inference
Top-3 Key Evidence (with URLs):
1. [Evidence A] — S-grade — [Source](URL)
2. [Evidence B] — A-grade — [Source](URL)
3. [Evidence C] — B-grade — [Source](URL)
Core Uncertainties:
- [Uncertainty 1]
- [Uncertainty 2]
Evidence grades are decoration — they are the LIFEBLOOD of your conclusion.
| Grade | Definition | Examples | Usable? |
|---|---|---|---|
| S | Primary academic lit / Official stats / Raw data APIs / Authoritative market reports | Peer-reviewed papers, government statistics, exchange data | ✅ Standalone |
| A | Respected media / Professional analysis / Fully cited secondary sources | Reuters, Bloomberg, financial analyst reports | ✅ Needs ≥1 corroboration |
| B | Industry analysis / Forum discussions / Indirect data / Raw executive quotes | CEO statements (cross-verified across video titles), tech news | ✅ Needs ≥2 cross-references |
| C | Social media / Single samples / Non-professional reading / Snippet from search results | Individual blog posts, Reddit answers, single YouTube title | ⚠️ Leads only, cannot conclude |
| D | No source / Rumors / Obvious conflict of interest / User comment section | YouTube/Reddit comment section, anonymous forum posts | ❌ Never use as evidence |
Video titles = C-grade (weak lead starting point)
Comment section user posts = D-grade (unreliable by default)
Source URLs are mandatory, not optional
Better to say less than to fabricate
🔴 Security Constraint: This skill's self-modification is gated behind explicit user commands.
User provides feedback → default action: update memory only (no skill file change) User says "update the skill" / "commit this to the skill" / "add this to the workflow" → then execute skill_manage
This prevents: malicious input injection / accidental trigger during research / unconfirmed auto-modification
When the user provides improvement feedback:
memory(action='add', ...) records preferences and lessonsskill_manage calls, no SKILL.md modificationOnly execute skill_manage(patch) when the user explicitly says:
Common trigger scenarios:
| User feedback type | Record to memory | Upgrade to skill |
|---|---|---|
| Direction correction: "This sub-question isn't the point" | ✅ Default | When user says "update the skill accordingly" |
| Evidence standard: "This source isn't good enough" | ✅ Default | When user says "add this to the evidence discipline" |
| Format preference: "Too long / give me a short version first" | ✅ Default | When user says "save this format to the skill" |
| New scenario: "This isn't just fact-checking, it's data research" | ✅ Default | When user says "add this to trigger conditions" |
| Methodology: "You should plan before executing" | ✅ Default | When user says "add this to the workflow" |
| Recurring error (≥2 same class) | ✅ Default | When user says "add this to common pitfalls" |
| Scenario | Characteristics | Watch Out For |
|---|---|---|
| Fact-check | Verify a specific claim | Find primary source, watch for telephone game distortions |
| Trend analysis | Predict direction of a metric | Separate short-term noise from long-term trends, note data window |
| Comparison | Compare options | Ensure full dimension coverage, avoid survivorship bias |
| Causal analysis | Did A cause B? | Distinguish correlation from causation, watch for confounders |
| Consumer pricing/product research | Product pricing and storage strategy | First verify if the product is already launched! Check executive statements; find raw component cost data from market research firms |