Co-Pilot / 辅助式
更新于 a month ago

review-implementing

Mmhattingpete
0.2k
mhattingpete/claude-skills-marketplace/engineering-workflow-plugin/skills/review-implementing
80
Agent 评分

💡 摘要

一个能系统化解析代码审查反馈并将其转化为可执行任务和代码修改的AI代理技能。

🎯 适合人群

软件工程师技术负责人开源项目维护者DevOps工程师代码审查员

🤖 AI 吐槽:它本质上是一个为代码审查服务的花哨待办事项管理器,以电子表格般的热情将人类反馈变成机器人的检查清单。

安全分析低风险

该技能执行代码编辑和文件系统操作(Grep, Glob, Edit, Write)。主要风险是对源代码进行未经授权或错误的修改,可能引入漏洞或破坏功能。缓解措施:对指定“沙盒”目录之外的修改实施强制性的试运行或确认步骤,并与版本控制系统集成以允许轻松回滚。


name: review-implementing description: Process and implement code review feedback systematically. Use when user provides reviewer comments, PR feedback, code review notes, or asks to implement suggestions from reviews.

Review Feedback Implementation

Systematically process and implement changes based on code review feedback.

When to Use

  • Provides reviewer comments or feedback
  • Pastes PR review notes
  • Mentions implementing review suggestions
  • Says "address these comments" or "implement feedback"
  • Shares list of changes requested by reviewers

Systematic Workflow

1. Parse Reviewer Notes

Identify individual feedback items:

  • Split numbered lists (1., 2., etc.)
  • Handle bullet points or unnumbered feedback
  • Extract distinct change requests
  • Clarify ambiguous items before starting

2. Create Todo List

Use TodoWrite tool to create actionable tasks:

  • Each feedback item becomes one or more todos
  • Break down complex feedback into smaller tasks
  • Make tasks specific and measurable
  • Mark first task as in_progress before starting

Example:

- Add type hints to extract function
- Fix duplicate tag detection logic
- Update docstring in chain.py
- Add unit test for edge case

3. Implement Changes Systematically

For each todo item:

Locate relevant code:

  • Use Grep to search for functions/classes
  • Use Glob to find files by pattern
  • Read current implementation

Make changes:

  • Use Edit tool for modifications
  • Follow project conventions (CLAUDE.md)
  • Preserve existing functionality unless changing behavior

Verify changes:

  • Check syntax correctness
  • Run relevant tests if applicable
  • Ensure changes address reviewer's intent

Update status:

  • Mark todo as completed immediately after finishing
  • Move to next todo (only one in_progress at a time)

4. Handle Different Feedback Types

Code changes:

  • Use Edit tool for existing code
  • Follow type hint conventions (PEP 604/585)
  • Maintain consistent style

New features:

  • Create new files with Write tool if needed
  • Add corresponding tests
  • Update documentation

Documentation:

  • Update docstrings following project style
  • Modify markdown files as needed
  • Keep explanations concise

Tests:

  • Write tests as functions, not classes
  • Use descriptive names
  • Follow pytest conventions

Refactoring:

  • Preserve functionality
  • Improve code structure
  • Run tests to verify no regressions

5. Validation

After implementing changes:

  • Run affected tests
  • Check for linting errors: uv run ruff check
  • Verify changes don't break existing functionality

6. Communication

Keep user informed:

  • Update todo list in real-time
  • Ask for clarification on ambiguous feedback
  • Report blockers or challenges
  • Summarize changes at completion

Edge Cases

Conflicting feedback:

  • Ask user for guidance
  • Explain conflict clearly

Breaking changes required:

  • Notify user before implementing
  • Discuss impact and alternatives

Tests fail after changes:

  • Fix tests before marking todo complete
  • Ensure all related tests pass

Referenced code doesn't exist:

  • Ask user for clarification
  • Verify understanding before proceeding

Important Guidelines

  • Always use TodoWrite for tracking progress
  • Mark todos completed immediately after each item
  • Only one todo in_progress at any time
  • Don't batch completions - update status in real-time
  • Ask questions for unclear feedback
  • Run tests if changes affect tested code
  • Follow CLAUDE.md conventions for all code changes
  • Use conventional commits if creating commits afterward
五维分析
清晰度8/10
创新性6/10
实用性9/10
完整性9/10
可维护性8/10
优缺点分析

优点

  • 为处理反馈提供了结构化、可重复的工作流程。
  • 减少了跟踪和实施审查项时的人为错误。
  • 与现有工具(Grep, Edit)集成,实现上下文感知的修改。
  • 促进清晰的沟通和状态更新。

缺点

  • 可能难以处理需要人类判断的高度模糊或主观的反馈。
  • 依赖用户提供结构良好的初始输入。
  • 对于非常小、直接的审查修改可能显得过度设计。
  • 为琐碎的修复增加了流程开销。

相关技能

useful-ai-prompts

A
toolCo-Pilot / 辅助式
88/ 100

“一个提示的宝藏,但别指望它们为你写小说。”

fastmcp

A
toolCo-Pilot / 辅助式
86/ 100

“FastMCP:因为谁不喜欢在AI中添加一点复杂性呢?”

python-pro

A
toolCo-Pilot / 辅助式
86/ 100

“看起来很能打,但别让配置把人劝退。”

免责声明:本内容来源于 GitHub 开源项目,仅供展示和评分分析使用。

版权归原作者所有 mhattingpete.