scientific-critical-thinking

85
17
Source

Evaluate research rigor. Assess methodology, experimental design, statistical validity, biases, confounding, evidence quality (GRADE, Cochrane ROB), for critical analysis of scientific claims.

Install

mkdir -p .claude/skills/scientific-critical-thinking && curl -L -o skill.zip "https://mcp.directory/api/skills/download/358" && unzip -o skill.zip -d .claude/skills/scientific-critical-thinking && rm skill.zip

Installs to .claude/skills/scientific-critical-thinking

About this skill

Scientific Critical Thinking

Overview

Critical thinking is a systematic process for evaluating scientific rigor. Assess methodology, experimental design, statistical validity, biases, confounding, and evidence quality using GRADE and Cochrane ROB frameworks. Apply this skill for critical analysis of scientific claims.

When to Use This Skill

This skill should be used when:

  • Evaluating research methodology and experimental design
  • Assessing statistical validity and evidence quality
  • Identifying biases and confounding in studies
  • Reviewing scientific claims and conclusions
  • Conducting systematic reviews or meta-analyses
  • Applying GRADE or Cochrane risk of bias assessments
  • Providing critical analysis of research papers

Visual Enhancement with Scientific Schematics

When creating documents with this skill, always consider adding scientific diagrams and schematics to enhance visual communication.

If your document does not already contain schematics or diagrams:

  • Use the scientific-schematics skill to generate AI-powered publication-quality diagrams
  • Simply describe your desired diagram in natural language
  • Nano Banana Pro will automatically generate, review, and refine the schematic

For new documents: Scientific schematics should be generated by default to visually represent key concepts, workflows, architectures, or relationships described in the text.

How to generate schematics:

python scripts/generate_schematic.py "your diagram description" -o figures/output.png

The AI will automatically:

  • Create publication-quality images with proper formatting
  • Review and refine through multiple iterations
  • Ensure accessibility (colorblind-friendly, high contrast)
  • Save outputs in the figures/ directory

When to add schematics:

  • Critical thinking framework diagrams
  • Bias identification decision trees
  • Evidence quality assessment flowcharts
  • GRADE assessment methodology diagrams
  • Risk of bias evaluation frameworks
  • Validity assessment visualizations
  • Any complex concept that benefits from visualization

For detailed guidance on creating schematics, refer to the scientific-schematics skill documentation.


Core Capabilities

1. Methodology Critique

Evaluate research methodology for rigor, validity, and potential flaws.

Apply when:

  • Reviewing research papers
  • Assessing experimental designs
  • Evaluating study protocols
  • Planning new research

Evaluation framework:

  1. Study Design Assessment

    • Is the design appropriate for the research question?
    • Can the design support causal claims being made?
    • Are comparison groups appropriate and adequate?
    • Consider whether experimental, quasi-experimental, or observational design is justified
  2. Validity Analysis

    • Internal validity: Can we trust the causal inference?
      • Check randomization quality
      • Evaluate confounding control
      • Assess selection bias
      • Review attrition/dropout patterns
    • External validity: Do results generalize?
      • Evaluate sample representativeness
      • Consider ecological validity of setting
      • Assess whether conditions match target application
    • Construct validity: Do measures capture intended constructs?
      • Review measurement validation
      • Check operational definitions
      • Assess whether measures are direct or proxy
    • Statistical conclusion validity: Are statistical inferences sound?
      • Verify adequate power/sample size
      • Check assumption compliance
      • Evaluate test appropriateness
  3. Control and Blinding

    • Was randomization properly implemented (sequence generation, allocation concealment)?
    • Was blinding feasible and implemented (participants, providers, assessors)?
    • Are control conditions appropriate (placebo, active control, no treatment)?
    • Could performance or detection bias affect results?
  4. Measurement Quality

    • Are instruments validated and reliable?
    • Are measures objective when possible, or subjective with acknowledged limitations?
    • Is outcome assessment standardized?
    • Are multiple measures used to triangulate findings?

Reference: See references/scientific_method.md for detailed principles and references/experimental_design.md for comprehensive design checklist.

2. Bias Detection

Identify and evaluate potential sources of bias that could distort findings.

Apply when:

  • Reviewing published research
  • Designing new studies
  • Interpreting conflicting evidence
  • Assessing research quality

Systematic bias review:

  1. Cognitive Biases (Researcher)

    • Confirmation bias: Are only supporting findings highlighted?
    • HARKing: Were hypotheses stated a priori or formed after seeing results?
    • Publication bias: Are negative results missing from literature?
    • Cherry-picking: Is evidence selectively reported?
    • Check for preregistration and analysis plan transparency
  2. Selection Biases

    • Sampling bias: Is sample representative of target population?
    • Volunteer bias: Do participants self-select in systematic ways?
    • Attrition bias: Is dropout differential between groups?
    • Survivorship bias: Are only "survivors" visible in sample?
    • Examine participant flow diagrams and compare baseline characteristics
  3. Measurement Biases

    • Observer bias: Could expectations influence observations?
    • Recall bias: Are retrospective reports systematically inaccurate?
    • Social desirability: Are responses biased toward acceptability?
    • Instrument bias: Do measurement tools systematically err?
    • Evaluate blinding, validation, and measurement objectivity
  4. Analysis Biases

    • P-hacking: Were multiple analyses conducted until significance emerged?
    • Outcome switching: Were non-significant outcomes replaced with significant ones?
    • Selective reporting: Are all planned analyses reported?
    • Subgroup fishing: Were subgroup analyses conducted without correction?
    • Check for study registration and compare to published outcomes
  5. Confounding

    • What variables could affect both exposure and outcome?
    • Were confounders measured and controlled (statistically or by design)?
    • Could unmeasured confounding explain findings?
    • Are there plausible alternative explanations?

Reference: See references/common_biases.md for comprehensive bias taxonomy with detection and mitigation strategies.

3. Statistical Analysis Evaluation

Critically assess statistical methods, interpretation, and reporting.

Apply when:

  • Reviewing quantitative research
  • Evaluating data-driven claims
  • Assessing clinical trial results
  • Reviewing meta-analyses

Statistical review checklist:

  1. Sample Size and Power

    • Was a priori power analysis conducted?
    • Is sample adequate for detecting meaningful effects?
    • Is the study underpowered (common problem)?
    • Do significant results from small samples raise flags for inflated effect sizes?
  2. Statistical Tests

    • Are tests appropriate for data type and distribution?
    • Were test assumptions checked and met?
    • Are parametric tests justified, or should non-parametric alternatives be used?
    • Is the analysis matched to study design (e.g., paired vs. independent)?
  3. Multiple Comparisons

    • Were multiple hypotheses tested?
    • Was correction applied (Bonferroni, FDR, other)?
    • Are primary outcomes distinguished from secondary/exploratory?
    • Could findings be false positives from multiple testing?
  4. P-Value Interpretation

    • Are p-values interpreted correctly (probability of data if null is true)?
    • Is non-significance incorrectly interpreted as "no effect"?
    • Is statistical significance conflated with practical importance?
    • Are exact p-values reported, or only "p < .05"?
    • Is there suspicious clustering just below .05?
  5. Effect Sizes and Confidence Intervals

    • Are effect sizes reported alongside significance?
    • Are confidence intervals provided to show precision?
    • Is the effect size meaningful in practical terms?
    • Are standardized effect sizes interpreted with field-specific context?
  6. Missing Data

    • How much data is missing?
    • Is missing data mechanism considered (MCAR, MAR, MNAR)?
    • How is missing data handled (deletion, imputation, maximum likelihood)?
    • Could missing data bias results?
  7. Regression and Modeling

    • Is the model overfitted (too many predictors, no cross-validation)?
    • Are predictions made outside the data range (extrapolation)?
    • Are multicollinearity issues addressed?
    • Are model assumptions checked?
  8. Common Pitfalls

    • Correlation treated as causation
    • Ignoring regression to the mean
    • Base rate neglect
    • Texas sharpshooter fallacy (pattern finding in noise)
    • Simpson's paradox (confounding by subgroups)

Reference: See references/statistical_pitfalls.md for detailed pitfalls and correct practices.

4. Evidence Quality Assessment

Evaluate the strength and quality of evidence systematically.

Apply when:

  • Weighing evidence for decisions
  • Conducting literature reviews
  • Comparing conflicting findings
  • Determining confidence in conclusions

Evidence evaluation framework:

  1. Study Design Hierarchy

    • Systematic reviews/meta-analyses (highest for intervention effects)
    • Randomized controlled trials
    • Cohort studies
    • Case-control studies
    • Cross-sectional studies
    • Case series/reports
    • Expert opinion (lowest)

    Important: Higher-level designs aren't always better quality. A well-designed observational study can be stronger than a poorly-conducted RCT.

  2. Quality Within Design Type

    • Risk of bias assessment (use appropriate tool: Cochrane ROB, Newcastle-Ottawa, etc.)
    • Methodological rigor
    • Transparency and reporting completeness
    • Conflicts

Content truncated.

software-architecture

davila7

Guide for quality focused software architecture. This skill should be used when users want to write code, design architecture, analyze code, in any case that relates to software development.

473165

scroll-experience

davila7

Expert in building immersive scroll-driven experiences - parallax storytelling, scroll animations, interactive narratives, and cinematic web experiences. Like NY Times interactives, Apple product pages, and award-winning web experiences. Makes websites feel like experiences, not just pages. Use when: scroll animation, parallax, scroll storytelling, interactive story, cinematic website.

12580

planning-with-files

davila7

Implements Manus-style file-based planning for complex tasks. Creates task_plan.md, findings.md, and progress.md. Use when starting complex multi-step tasks, research projects, or any task requiring >5 tool calls.

7968

humanizer

davila7

Remove signs of AI-generated writing from text. Use when editing or reviewing text to make it sound more natural and human-written. Based on Wikipedia's comprehensive "Signs of AI writing" guide. Detects and fixes patterns including: inflated symbolism, promotional language, superficial -ing analyses, vague attributions, em dash overuse, rule of three, AI vocabulary words, negative parallelisms, and excessive conjunctive phrases. Credits: Original skill by @blader - https://github.com/blader/humanizer

10352

game-development

davila7

Game development orchestrator. Routes to platform-specific skills based on project needs.

14749

2d-games

davila7

2D game development principles. Sprites, tilemaps, physics, camera.

12944

You might also like

flutter-development

aj-geddes

Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.

1,5751,370

ui-ux-pro-max

nextlevelbuilder

"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."

1,1181,191

drawio-diagrams-enhanced

jgtolentino

Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.

1,4191,110

godot

bfollington

This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.

1,199751

nano-banana-pro

garg-aayush

Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.

1,159685

pdf-to-markdown

aliceisjustplaying

Convert entire PDF documents to clean, structured Markdown for full context loading. Use this skill when the user wants to extract ALL text from a PDF into context (not grep/search), when discussing or analyzing PDF content in full, when the user mentions "load the whole PDF", "bring the PDF into context", "read the entire PDF", or when partial extraction/grepping would miss important context. This is the preferred method for PDF text extraction over page-by-page or grep approaches.

1,324620

Stay ahead of the MCP ecosystem

Get weekly updates on new skills and servers.