Web Research

Web Research

mzxrai

Performs Google searches and scrapes webpage content to bring real-time web information into your conversations.

Research topics using Google search and web scraping.

2921,895 views74Local (stdio)

What it does

  • Search Google for current information
  • Extract content from webpages
  • Take screenshots of web pages
  • Track research sessions with visited pages and queries

Best for

Getting up-to-date information on current eventsResearch projects requiring recent dataContent creators needing current market insights
Includes guided research prompt templateZero setup with npx installation

About Web Research

Web Research is a community-built MCP server published by mzxrai that provides AI assistants with tools and capabilities via the Model Context Protocol. Research any topic fast with Web Research – powerful web scraping tools and advanced Google dorking for results you can It is categorized under search web, browser automation.

How to install

You can install Web Research in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

Web Research is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

MCP Web Research Server

A Model Context Protocol (MCP) server for web research.

Bring real-time info into Claude and easily research any topic.

Features

  • Google search integration
  • Webpage content extraction
  • Research session tracking (list of visited pages, search queries, etc.)
  • Screenshot capture

Prerequisites

Installation

First, ensure you've downloaded and installed the Claude Desktop app and you have npm installed.

Next, add this entry to your claude_desktop_config.json (on Mac, found at ~/Library/Application\ Support/Claude/claude_desktop_config.json):

{
  "mcpServers": {
    "webresearch": {
      "command": "npx",
      "args": ["-y", "@mzxrai/mcp-webresearch@latest"]
    }
  }
}

This config allows Claude Desktop to automatically start the web research MCP server when needed.

Usage

Simply start a chat with Claude and send a prompt that would benefit from web research. If you'd like a prebuilt prompt customized for deeper web research, you can use the agentic-research prompt that we provide through this package. Access that prompt in Claude Desktop by clicking the Paperclip icon in the chat input and then selecting Choose an integrationwebresearchagentic-research.

Example screenshot of web research

Tools

  1. search_google

    • Performs Google searches and extracts results
    • Arguments: { query: string }
  2. visit_page

    • Visits a webpage and extracts its content
    • Arguments: { url: string, takeScreenshot?: boolean }
  3. take_screenshot

    • Takes a screenshot of the current page
    • No arguments required

Prompts

agentic-research

A guided research prompt that helps Claude conduct thorough web research. The prompt instructs Claude to:

  • Start with broad searches to understand the topic landscape
  • Prioritize high-quality, authoritative sources
  • Iteratively refine the research direction based on findings
  • Keep you informed and let you guide the research interactively
  • Always cite sources with URLs

Resources

We expose two things as MCP resources: (1) captured webpage screenshots, and (2) the research session.

Screenshots

When you take a screenshot, it's saved as an MCP resource. You can access captured screenshots in Claude Desktop via the Paperclip icon.

Research Session

The server maintains a research session that includes:

  • Search queries
  • Visited pages
  • Extracted content
  • Screenshots
  • Timestamps

Suggestions

For the best results, if you choose not to use the agentic-research prompt when doing your research, it may be helpful to suggest high-quality sources for Claude to use when researching general topics. For example, you could prompt news today from reuters or AP instead of news today.

Problems

This is very much pre-alpha code. And it is also AIGC, so expect bugs.

If you run into issues, it may be helpful to check Claude Desktop's MCP logs:

tail -n 20 -f ~/Library/Logs/Claude/mcp*.log

Development

# Install dependencies
pnpm install

# Build the project
pnpm build

# Watch for changes
pnpm watch

# Run in development mode
pnpm dev

Requirements

  • Node.js >= 18
  • Playwright (automatically installed as a dependency)

Verified Platforms

  • macOS
  • Linux

License

MIT

Author

mzxrai

Alternatives

Related Skills

Browse all skills
browser-automation

Automate web browser interactions using natural language via CLI commands. Use when the user asks to browse websites, navigate web pages, extract data from websites, take screenshots, fill forms, click buttons, or interact with web applications. Triggers include "browse", "navigate to", "go to website", "extract data from webpage", "screenshot", "web scraping", "fill out form", "click on", "search for on the web". When taking actions be as specific as possible.

16
aluvia-web-unblock

Unblock websites and bypass CAPTCHAs and 403 errors using Aluvia mobile proxies. Enables web search and content extraction without browser automation.

0
aluvia-web-proxy

Unblock websites and bypass CAPTCHAs and 403 errors using Aluvia mobile proxies. Enables web search and content extraction without browser automation.

0
competitor-scan

Research best-in-class products using Browser MCP and WebSearch

0
dev-browser

Browser automation with persistent page state. Use when users ask to navigate websites, fill forms, take screenshots, extract web data, test web apps, or automate browser workflows. Trigger phrases include "go to [url]", "click on", "fill out the form", "take a screenshot", "scrape", "automate", "test the website", "log into", or any browser interaction request.

29
chrome-devtools

Browser automation, debugging, and performance analysis using Puppeteer CLI scripts. Use for automating browsers, taking screenshots, analyzing performance, monitoring network traffic, web scraping, form automation, and JavaScript debugging.

21