coderabbit

This skill should be used when the user asks to "run coderabbit review", "coderabbit review", "review with coderabbit", "triage coderabbit findings", "check code with coderabbit", or mentions running CodeRabbit CLI review, triaging CodeRabbit output, or evaluating CodeRabbit suggestions on local changes.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "coderabbit" with this command: npx skills add paulrberg/agent-skills/paulrberg-agent-skills-coderabbit

CodeRabbit Review

Run a CodeRabbit CLI review on local changes and triage the output by verifying each finding against the actual code. Classify findings as Valid or False Positive, assign severity, and produce a prioritized fix plan. Not every CodeRabbit suggestion deserves a code change — the goal is to separate signal from noise.

Arguments

Parse $ARGUMENTS for optional flags:

  • --base <branch> — base branch for diff comparison
  • --base-commit <commit> — base commit on current branch for comparison (mutually exclusive with --base)
  • --type <type> — review scope: committed, uncommitted, or all (default: all)
  • --config <files...> — additional instruction files to pass to CodeRabbit

All flags are optional. If neither --base nor --base-commit is provided, auto-detect the base branch:

  1. Try git rev-parse --abbrev-ref @{upstream} (strip remote prefix)
  2. Fall back to git symbolic-ref refs/remotes/origin/HEAD (strip remote prefix)
  3. Fall back to main, then master

Verify the resolved ref exists with git rev-parse --verify.

Prerequisites

Run these checks in order. Stop at the first failure.

1) CodeRabbit CLI Installed

command -v coderabbit

If missing, stop and tell the user to install:

curl -fsSL https://cli.coderabbit.ai/install.sh | sh

2) Authentication

coderabbit auth status

If the exit code indicates unauthenticated or the output does not show a logged-in state, stop and tell the user to authenticate:

coderabbit auth login

Workflow

1) Run CodeRabbit Review

Construct the command with mandatory flags --plain --no-color plus any user-provided flags:

coderabbit review --plain --no-color [--base <branch>] [--base-commit <commit>] [--type <type>] [--config <files...>]

Capture the full output. If the CLI exits non-zero, report the error and stop.

2) Parse and Group

Parse the CLI output into individual findings. Group by file path. For each finding, categorize into:

  • AI-actionable: concrete code suggestion with a clear, specific fix
  • Nitpick: style or naming preference without functional impact
  • Informational: explanation or context without a specific suggestion

3) Classify Each Finding

For each AI-actionable finding, perform a thorough assessment:

  1. Read the actual code at the referenced file and line number. Do not rely solely on the CLI's quoted snippet — it may be outdated or truncated.
  2. Check project conventions by examining linter configs (ESLint, Prettier, Biome, Ruff, etc.), existing patterns in the codebase, and any style guides or CONTRIBUTING.md files.
  3. Review broader context by reading surrounding functions, the module's purpose, and related tests to understand whether the suggestion fits.
  4. Assess applicability — does this suggestion make sense for this specific codebase, or is it generic advice that doesn't apply here?
  5. Evaluate severity based on real impact to the system, not just theoretical concern.

Assign one of two classifications:

  • Valid with severity: CRITICAL > HIGH > MEDIUM > LOW
    • CRITICAL: exploitable security flaw, data loss path, or outage risk on critical paths
    • HIGH: logic defect or performance failure that can break core behavior
    • MEDIUM: maintainability or reliability issue likely to cause near-term defects
    • LOW: localized clarity, style, or documentation improvements
  • False Positive with a specific reason:
    • Incorrect assumption about the code's behavior or purpose
    • Project convention mismatch (suggestion contradicts established patterns)
    • Already handled elsewhere in the codebase
    • Out of scope for the current diff's intent

3b) Confirm Ambiguous Classifications

After classifying all findings, identify those where confidence is Medium or Low and the code is on a critical path (auth, payments, data integrity, core business logic). For each such finding, use AskUserQuestion to ask the user:

  • header: Truncated filename (max 12 chars)
  • question: "CodeRabbit flagged <file>:<line><one-line summary of suggestion>. Is this valid?"
  • options:
    • "Valid — fix it" with description of what the fix entails
    • "False positive — skip" with description of why it might not apply
    • "Needs context — defer" with description that it will be added to Residual Risks
  • multiSelect: false

Batch up to 4 questions per AskUserQuestion call to minimize interruptions. Update classifications based on user responses before generating the fix plan.

Skip this step when all medium/low-confidence findings are on non-critical paths — classify those using your best judgment and note the confidence level in the report.

4) Generate Fix Plan

For each valid finding, ordered by severity (CRITICAL first, LOW last), produce a structured entry:

FieldContent
Locationfile:line
CodeRabbit suggestionBrief summary of what was suggested
AssessmentWhy this is valid and what the real impact is
Proposed fixConcrete code change description
ConfidenceHigh / Medium / Low

Confidence reflects certainty that the fix is correct and safe to apply:

  • High: clear defect with an obvious fix, no ambiguity
  • Medium: likely correct but depends on runtime behavior or external state not fully visible
  • Low: plausible improvement but may have side effects or require domain knowledge to validate

5) Report

Output a structured report with the following sections:

Scope: base ref, review type, number of files in diff.

Summary: N findings from CLI, N classified as valid, N as false positive, N nitpicks skipped.

Valid Findings (grouped by severity):

For each severity level present, list findings with location, CodeRabbit's original suggestion, your assessment, and the proposed fix. CRITICAL and HIGH findings should include explicit reasoning about blast radius and failure modes.

False Positives:

For each dismissed finding, list the location, what CodeRabbit suggested, and the specific rationale for dismissal.

Fix Plan:

Ordered implementation steps for all valid findings. Group related fixes that touch the same file. Note any fixes that should be applied together to avoid intermediate broken states.

Nitpicks (skipped):

Brief list of nitpick findings that were excluded from triage — mention location and what was suggested, but no fix plan entry.

Residual Risks:

Flag anything that needs human judgment or falls beyond automated triage — ambiguous intent, architectural concerns, suggestions that require product decisions, or findings where classification confidence is low on critical-path code.

Stop Conditions

Stop and ask for direction when:

  • coderabbit CLI is not installed or not authenticated.
  • The CLI exits with an error.
  • The resolved base ref does not exist.
  • No findings in the CLI output (report clean and exit).
  • A suggestion requires architectural changes beyond the current diff's scope.
  • Classification confidence is low for code on critical paths (auth, payments, data integrity).

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

code-simplify

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

code-review

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

cli-gh

No summary provided by upstream source.

Repository SourceNeeds Review