zoom-meeting-assistance-rtms-unofficial-community

0
0
Source

Zoom RTMS Meeting Assistant — start on-demand to capture meeting audio, video, transcript, screenshare, and chat via Zoom Real-Time Media Streams. Handles meeting.rtms_started and meeting.rtms_stopped webhook events. Provides AI-powered dialog suggestions, sentiment analysis, and live summaries with WhatsApp notifications. Use when a Zoom RTMS webhook fires or the user asks to record/analyze a meeting.

Install

mkdir -p .claude/skills/zoom-meeting-assistance-rtms-unofficial-community && curl -L -o skill.zip "https://mcp.directory/api/skills/download/9150" && unzip -o skill.zip -d .claude/skills/zoom-meeting-assistance-rtms-unofficial-community && rm skill.zip

Installs to .claude/skills/zoom-meeting-assistance-rtms-unofficial-community

About this skill

Zoom RTMS Meeting Assistant

Headless capture service for Zoom meetings using Real-Time Media Streams (RTMS). Receives webhook events, connects to RTMS WebSockets, records all media, and runs AI analysis via OpenClaw.

Webhook Events Handled

This skill processes two Zoom webhook events:

  • meeting.rtms_started — Zoom sends this when RTMS is activated for a meeting. Contains server_urls, rtms_stream_id, and meeting_uuid needed to connect to the RTMS WebSocket.
  • meeting.rtms_stopped — Zoom sends this when RTMS ends (meeting ended or RTMS disabled). Triggers cleanup: closes WebSocket connections, generates screenshare PDF, sends summary notification.

Webhook Dependency

This skill needs a public webhook endpoint to receive these events from Zoom.

Preferred: Use the ngrok-unofficial-webhook-skill (skills/ngrok-unofficial-webhook-skill). It auto-discovers this skill via webhookEvents in skill.json, notifies the user, and offers to route events here.

Other webhook solutions (e.g. custom servers, cloud functions) will work but require additional integration to forward payloads to this service.

Prerequisites

cd skills/zoom-meeting-assistance-rtms-unofficial-community
npm install

Requires ffmpeg for post-meeting media conversion.

Environment Variables

Set these in the skill's .env file:

Required:

  • ZOOM_SECRET_TOKEN — Zoom webhook secret token
  • ZOOM_CLIENT_ID — Zoom app Client ID
  • ZOOM_CLIENT_SECRET — Zoom app Client Secret

Optional:

  • PORT — Server port (default: 3000)
  • AI_PROCESSING_INTERVAL_MS — AI analysis frequency in ms (default: 30000)
  • AI_FUNCTION_STAGGER_MS — Delay between AI calls in ms (default: 5000)
  • AUDIO_DATA_OPT1 = mixed stream, 2 = multi-stream (default: 2)
  • OPENCLAW_NOTIFY_CHANNEL — Notification channel (default: whatsapp)
  • OPENCLAW_NOTIFY_TARGET — Phone number / target for notifications

Starting the Service

cd skills/zoom-meeting-assistance-rtms-unofficial-community
node index.js

This starts an Express server listening for Zoom webhook events on PORT.

⚠️ Important: Before forwarding webhooks to this service, always check if it's running:

# Check if service is listening on port 3000
lsof -i :3000

If nothing is returned, start the service first before forwarding any webhook events.

Typical flow:

  1. Start the server as a background process
  2. Zoom sends meeting.rtms_started webhook → service connects to RTMS WebSocket
  3. Media streams in real-time: audio, video, transcript, screenshare, chat
  4. AI processing runs periodically (dialog suggestions, sentiment, summary)
  5. meeting.rtms_stopped → service closes connections, generates screenshare PDF

Recorded Data

All recordings are stored organized by date:

skills/zoom-meeting-assistance-rtms-unofficial-community/recordings/YYYY/MM/DD/{streamId}/

Each stream folder contains:

FileContentSearchable
metadata.jsonMeeting metadata (UUID, stream ID, operator, start time)
transcript.txtPlain text transcript with timestamps and speaker names✅ Best for searching — grep-friendly, one line per utterance
transcript.vttVTT format transcript with timing cues
transcript.srtSRT format transcript
events.logParticipant join/leave, active speaker changes (JSON lines)
chat.txtChat messages with timestamps
ai_summary.mdAI-generated meeting summary (markdown)✅ Key document — read this first for meeting overview
ai_dialog.jsonAI dialog suggestions
ai_sentiment.jsonSentiment analysis per participant
mixedaudio.rawMixed audio stream (raw PCM)❌ Binary
activespeakervideo.h264Active speaker video (raw H.264)❌ Binary
processed/screenshare.pdfDeduplicated screenshare frames as PDF❌ Binary

All summaries are also copied to a central folder for easy access:

skills/zoom-meeting-assistance-rtms-unofficial-community/summaries/summary_YYYY-MM-DDTHH-MM-SS_{streamId}.md

Searching & Querying Past Meetings

To find and review past meeting data:

# List all recorded meetings by date
ls -R recordings/

# List meetings for a specific date
ls recordings/2026/01/28/

# Search across all transcripts for a keyword
grep -rl "keyword" recordings/*/*/*/*/transcript.txt

# Search for what a specific person said
grep "Chun Siong Tan" recordings/*/*/*/*/transcript.txt

# Read a meeting summary
cat recordings/YYYY/MM/DD/<streamId>/ai_summary.md

# Search summaries for a topic
grep -rl "topic" recordings/*/*/*/*/ai_summary.md

# Check who attended a meeting
cat recordings/YYYY/MM/DD/<streamId>/events.log

# Get sentiment for a meeting
cat recordings/YYYY/MM/DD/<streamId>/ai_sentiment.json

The .txt, .md, .json, and .log files are all text-based and searchable. Start with ai_summary.md for a quick overview, then drill into transcript.txt for specific quotes or details.

API Endpoints

# Toggle WhatsApp notifications on/off
curl -X POST http://localhost:3000/api/notify-toggle -H "Content-Type: application/json" -d '{"enabled": false}'

# Check notification status
curl http://localhost:3000/api/notify-toggle

Post-Meeting Processing

When meeting.rtms_stopped fires, the service automatically:

  1. Generates PDF from screenshare images
  2. Converts mixedaudio.rawmixedaudio.wav
  3. Converts activespeakervideo.h264activespeakervideo.mp4
  4. Muxes mixed audio + active speaker video into final_output.mp4

Manual conversion scripts are available but note that auto-conversion runs on meeting end, so manual re-runs are rarely needed.

Reading Meeting Data

After or during a meeting, read files from recordings/YYYY/MM/DD/{streamId}/:

# List recorded meetings by date
ls -R recordings/

# Read transcript
cat recordings/YYYY/MM/DD/<streamId>/transcript.txt

# Read AI summary
cat recordings/YYYY/MM/DD/<streamId>/ai_summary.md

# Read sentiment analysis
cat recordings/YYYY/MM/DD/<streamId>/ai_sentiment.json

Prompt Customization

Want different summary styles or analysis? Customize the AI prompts to fit your needs!

Edit these files to change AI behavior:

FilePurposeExample Customizations
summary_prompt.mdMeeting summary generationBullet points vs prose, focus areas, length
query_prompt.mdQuery response formattingResponse style, detail level
query_prompt_current_meeting.mdReal-time meeting analysisWhat to highlight during meetings
query_prompt_dialog_suggestions.mdDialog suggestion styleFormal vs casual, suggestion count
query_prompt_sentiment_analysis.mdSentiment scoring logicCustom sentiment categories, thresholds

Tip: Back up the originals before editing, so you can revert if needed.

You might also like

flutter-development

aj-geddes

Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.

1,4071,302

drawio-diagrams-enhanced

jgtolentino

Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.

1,2201,024

ui-ux-pro-max

nextlevelbuilder

"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."

9001,013

godot

bfollington

This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.

958658

nano-banana-pro

garg-aayush

Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.

970608

pdf-to-markdown

aliceisjustplaying

Convert entire PDF documents to clean, structured Markdown for full context loading. Use this skill when the user wants to extract ALL text from a PDF into context (not grep/search), when discussing or analyzing PDF content in full, when the user mentions "load the whole PDF", "bring the PDF into context", "read the entire PDF", or when partial extraction/grepping would miss important context. This is the preferred method for PDF text extraction over page-by-page or grep approaches.

1,033496

Stay ahead of the MCP ecosystem

Get weekly updates on new skills and servers.