coderabbit-migration-deep-dive
Execute CodeRabbit major re-architecture and migration strategies with strangler fig pattern. Use when migrating to or from CodeRabbit, performing major version upgrades, or re-platforming existing integrations to CodeRabbit. Trigger with phrases like "migrate coderabbit", "coderabbit migration", "switch to coderabbit", "coderabbit replatform", "coderabbit upgrade major".
Install
mkdir -p .claude/skills/coderabbit-migration-deep-dive && curl -L -o skill.zip "https://mcp.directory/api/skills/download/4674" && unzip -o skill.zip -d .claude/skills/coderabbit-migration-deep-dive && rm skill.zipInstalls to .claude/skills/coderabbit-migration-deep-dive
About this skill
CodeRabbit Migration Deep Dive
Overview
Comprehensive guide for migrating to CodeRabbit from other AI code review tools (Codacy, SonarCloud, DeepSource, Sourcery) or from manual-only code review. Covers assessment, phased rollout, configuration transfer, team buy-in, and measuring success.
Prerequisites
- GitHub/GitLab organization admin access
- Inventory of current review tools and their configurations
- Understanding of team review workflows
- Budget approval for CodeRabbit seats
Migration Types
| From | Complexity | Duration | Key Challenge |
|---|---|---|---|
| Manual-only reviews | Low | 1-2 weeks | Team adoption |
| Codacy / SonarCloud | Medium | 2-3 weeks | Rule translation |
| DeepSource / Sourcery | Medium | 2-3 weeks | Config migration |
| Custom review bots | High | 3-4 weeks | Workflow redesign |
| Multiple tools | High | 4-6 weeks | Consolidation |
Instructions
Step 1: Assess Current State
set -euo pipefail
ORG="${1:-your-org}"
echo "=== Code Review Tool Assessment ==="
# Check for existing review tools
echo "--- Installed GitHub Apps ---"
gh api "orgs/$ORG/installations" --jq '.installations[] | "\(.app_slug) (ID: \(.id))"' 2>/dev/null
echo ""
echo "--- Review Tool Config Files ---"
for REPO in $(gh repo list "$ORG" --limit 20 --json name --jq '.[].name'); do
# Check for common review tool configs
for CONFIG in ".codacy.yml" "sonar-project.properties" ".deepsource.toml" ".sourcery.yaml" ".coderabbit.yaml"; do
EXISTS=$(gh api "repos/$ORG/$REPO/contents/$CONFIG" --jq '.name' 2>/dev/null || echo "")
if [ -n "$EXISTS" ]; then
echo " $REPO: $CONFIG"
fi
done
done
Step 2: Map Review Rules to CodeRabbit Path Instructions
# Common rule translations:
# Codacy / SonarCloud "code smells" → CodeRabbit path_instructions
# Before (Codacy):
# rules:
# - id: "javascript/complexity"
# - id: "javascript/error-handling"
#
# After (CodeRabbit):
reviews:
path_instructions:
- path: "src/**/*.ts"
instructions: |
Check for:
- Functions with cyclomatic complexity > 10 (suggest refactoring)
- Missing error handling in async operations
- Empty catch blocks
- Unused variables and imports
# DeepSource "analyzer" → CodeRabbit path_instructions
# Before (DeepSource):
# analyzers:
# - name: javascript
# enabled: true
# meta:
# plugins: [react]
#
# After (CodeRabbit):
- path: "src/components/**"
instructions: |
React-specific checks:
- No conditional hooks
- Proper cleanup in useEffect
- Memoization for expensive computations
- Accessibility (aria labels, keyboard navigation)
# Sourcery "refactoring" → CodeRabbit path_instructions
# Before (Sourcery):
# refactor:
# skip: [dont-import-test-modules]
#
# After (CodeRabbit):
- path: "**/*.py"
instructions: |
Python best practices:
- Suggest list comprehensions over manual loops where appropriate
- Flag mutable default arguments
- Check for proper context manager usage
Step 3: Phase 1 -- Parallel Run (Week 1-2)
# Run CodeRabbit alongside existing tool for comparison
# .coderabbit.yaml - Start with non-blocking mode
reviews:
profile: "chill" # Fewer comments during evaluation
request_changes_workflow: false # Don't block merges
high_level_summary: true # Show walkthrough for evaluation
auto_review:
enabled: true
drafts: false
base_branches: [main, develop]
path_filters:
- "!**/*.lock"
- "!**/*.snap"
- "!dist/**"
- "!vendor/**"
chat:
auto_reply: true
# During parallel run, track:
1. Comment quality: Are CodeRabbit comments actionable?
2. Coverage: Does it catch what the old tool catches?
3. Speed: Is review posted before human reviewers start?
4. Noise: Are there many false positives?
5. Team reaction: Do developers find it helpful?
Step 4: Phase 2 -- Primary Tool (Week 3-4)
# After successful parallel run, make CodeRabbit primary
# .coderabbit.yaml - Enable full features
reviews:
profile: "assertive" # Balanced feedback
request_changes_workflow: true # Now blocking
high_level_summary: true
sequence_diagrams: true
auto_review:
enabled: true
drafts: false
base_branches: [main, develop]
path_instructions:
# Transfer your best rules from the old tool
- path: "src/api/**"
instructions: |
Review for: input validation, proper HTTP status codes,
auth middleware usage, error response format.
- path: "src/db/**"
instructions: |
Review for: parameterized queries, transaction boundaries,
connection cleanup, index usage. Flag N+1 patterns.
- path: "**/*.test.*"
instructions: |
Review for: assertion completeness, edge cases, async handling.
Do NOT comment on test naming or import order.
# Keep exclusions from old tool
path_filters:
- "!**/*.lock"
- "!**/*.snap"
- "!**/generated/**"
- "!dist/**"
- "!vendor/**"
Step 5: Phase 3 -- Decommission Old Tool (Week 4-6)
set -euo pipefail
ORG="${1:-your-org}"
echo "=== Old Tool Decommission Checklist ==="
# 1. Remove old tool config files
echo "--- Config Files to Remove ---"
for REPO in $(gh repo list "$ORG" --limit 50 --json name --jq '.[].name'); do
for CONFIG in ".codacy.yml" "sonar-project.properties" ".deepsource.toml" ".sourcery.yaml"; do
EXISTS=$(gh api "repos/$ORG/$REPO/contents/$CONFIG" --jq '.name' 2>/dev/null || echo "")
if [ -n "$EXISTS" ]; then
echo " rm $REPO/$CONFIG"
fi
done
done
echo ""
echo "--- Steps ---"
echo "1. Remove old tool GitHub App from org settings"
echo "2. Delete old tool config files from repos"
echo "3. Update branch protection rules (replace old check with coderabbitai)"
echo "4. Cancel old tool subscription"
echo "5. Update team documentation and onboarding guides"
Step 6: Measure Migration Success
set -euo pipefail
ORG="${1:-your-org}"
REPO="${2:-your-repo}"
echo "=== CodeRabbit Adoption Metrics ==="
# Review coverage
TOTAL=0
REVIEWED=0
for PR_NUM in $(gh api "repos/$ORG/$REPO/pulls?state=closed&per_page=30" --jq '.[].number'); do
TOTAL=$((TOTAL + 1))
CR=$(gh api "repos/$ORG/$REPO/pulls/$PR_NUM/reviews" \
--jq '[.[] | select(.user.login=="coderabbitai[bot]")] | length' 2>/dev/null || echo "0")
[ "$CR" -gt 0 ] && REVIEWED=$((REVIEWED + 1))
done
echo "Review coverage: $REVIEWED/$TOTAL PRs ($(( REVIEWED * 100 / (TOTAL > 0 ? TOTAL : 1) ))%)"
echo ""
echo "Target metrics:"
echo " - Coverage > 90%: CodeRabbit reviewing most PRs"
echo " - Time-to-review < 5 min: Fast feedback loop"
echo " - Team satisfaction: Survey developers after 2 weeks"
Output
- Current review tool assessment completed
- Rule translation from old tool to CodeRabbit path_instructions
- Phased migration plan executed
- Old tool decommissioned
- Adoption metrics measured
Error Handling
| Issue | Cause | Solution |
|---|---|---|
| Old tool conflicts with CodeRabbit | Both posting reviews | Run parallel briefly, then disable old tool |
| Rules don't translate 1:1 | Different analysis approaches | Focus on intent, not exact rule matching |
| Team prefers old tool | Familiarity bias | Run parallel for 2 weeks, compare results |
| Branch protection breaks | Old check name removed | Update to coderabbitai check name |
| Higher seat cost than old tool | Per-seat vs per-repo pricing | Scope repos to reduce seat count |
Resources
- CodeRabbit vs Alternatives
- CodeRabbit Configuration Reference
- Migration from Codacy
- CodeRabbit Path Instructions
Next Steps
For ongoing configuration tuning, see coderabbit-core-workflow-b.
More by jeremylongshore
View all skills by jeremylongshore →You might also like
flutter-development
aj-geddes
Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.
drawio-diagrams-enhanced
jgtolentino
Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.
ui-ux-pro-max
nextlevelbuilder
"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."
godot
bfollington
This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.
nano-banana-pro
garg-aayush
Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.
fastapi-templates
wshobson
Create production-ready FastAPI projects with async patterns, dependency injection, and comprehensive error handling. Use when building new FastAPI applications or setting up backend API projects.
Related MCP Servers
Browse all serversConnect Blender to Claude AI for seamless 3D modeling. Use AI 3D model generator tools for faster, intuitive, interactiv
Terminal control, file system search, and diff-based file editing for Claude and other AI assistants. Execute shell comm
Official Laravel-focused MCP server for augmenting AI-powered local development. Provides deep context about your Larave
Securely join MySQL databases with Read MySQL for read-only query access and in-depth data analysis.
AppleScript MCP server lets AI execute apple script on macOS, accessing Notes, Calendar, Contacts, Messages & Finder via
AIPo Labs — dynamic search and execute any tools available on ACI.dev for fast, flexible AI-powered workflows.
Stay ahead of the MCP ecosystem
Get weekly updates on new skills and servers.