Co-Pilot / 辅助式
更新于 a month ago

ai-research-skills

ZzechenzhangAGI
1.2k
zechenzhangagi/ai-research-skills
80
Agent 评分

💡 摘要

该库提供了一整套全面的AI研究工程技能,用于构建和部署AI代理。

🎯 适合人群

AI研究人员数据科学家机器学习工程师AI开发者学术机构

🤖 AI 吐槽:看起来很能打,但别让配置把人劝退。

安全分析中风险

自述文件没有明确提到安全措施,这引发了对潜在风险的担忧,例如依赖漏洞和未经授权的访问。实施依赖管理工具和定期安全审计可以减轻这些风险。

AI Research Engineering Skills Library

The most comprehensive open-source library of AI research engineering skills for AI agents

License: MIT Blog Post Demo

77 Skills Powering AI Research in 2026

| | | | |:---:|:---:|:---:| | Model Architecture (5) | Fine-Tuning (4) | Post-Training (4) | | Distributed Training (5) | Optimization (6) | Inference (4) | | Tokenization (2) | Data Processing (2) | Evaluation (3) | | Safety & Alignment (3) | Agents (4) | RAG (5) | | Multimodal (7) | Prompt Engineering (4) | MLOps (3) | | Observability (2) | Infrastructure (3) | Mech Interp (4) | | Emerging Techniques (6) | ML Paper Writing (1) | |


Table of Contents

Our Mission

We provide the layer of Engineering Ability that enable your coding agent to write and conduct AI research experiments, including preparing datasets, executing training pipelines, deploying models, and building your AI agents.

Path Towards AI Research Agent

Modern AI research requires mastering dozens of specialized tools and frameworks. AI Researchers spend more time debugging infrastructure than testing hypotheses—slowing the pace of scientific discovery. We provide a comprehensive library of expert-level research engineering skills that enable AI agents to autonomously implement and execute different stages of AI research experiments—from data preparation and model training to evaluation and deployment.

  • Specialized Expertise - Each skill provides deep, production-ready knowledge of a specific framework (Megatron-LM, vLLM, TRL, etc.)
  • End-to-End Coverage - 77 skills spanning model architecture, tokenization, fine-tuning, mechanistic interpretability, data processing, post-training, distributed training, optimization, evaluation, inference, infrastructure, agents, RAG, multimodal, prompt engineering, MLOps, observability, emerging techniques, and ML paper writing
  • Research-Grade Quality - Documentation sourced from official repos, real GitHub issues, and battle-tested production workflows

Available AI Research Engineering Skills

Quality over quantity: Each skill provides comprehensive, expert-level guidance with real code examples, troubleshooting guides, and production-ready workflows.

📦 Install from Claude Code Marketplace

Install skill categories directly using the Claude Code CLI:

# Add the marketplace /plugin marketplace add zechenzhangAGI/AI-research-SKILLs # Install by category (20 categories available) /plugin install fine-tuning@ai-research-skills # Axolotl, LLaMA-Factory, PEFT, Unsloth /plugin install post-training@ai-research-skills # TRL, GRPO, OpenRLHF, SimPO /plugin install inference-serving@ai-research-skills # vLLM, TensorRT-LLM, llama.cpp, SGLang /plugin install distributed-training@ai-research-skills /plugin install optimization@ai-research-skills

All 20 Categories: | Category | Install Command | Skills Included | |----------|-----------------|-----------------| | Model Architecture | model-architecture@ai-research-skills | LitGPT, Mamba, NanoGPT, RWKV | | Tokenization | tokenization@ai-research-skills | HuggingFace Tokenizers, SentencePiece | | Fine-Tuning | fine-tuning@ai-research-skills | Axolotl, LLaMA-Factory, PEFT, Unsloth | | Mech Interp | mechanistic-interpretability@ai-research-skills | TransformerLens, SAELens, pyvene, nnsight | | Data Processing | data-processing@ai-research-skills | NeMo Curator, Ray Data | | Post-Training | post-training@ai-research-skills | TRL, GRPO, OpenRLHF, SimPO | | Safety | safety-alignment@ai-research-skills | Constitutional AI, LlamaGuard, NeMo Guardrails | | Distributed | distributed-training@ai-research-skills | DeepSpeed, FSDP, Accelerate, Megatron, Lightning, Ray Train | | Infrastructure | infrastructure@ai-research-skills | Modal, Lambda Labs, SkyPilot | | Optimization | optimization@ai-research-skills | Flash Attention, bitsandbytes, GPTQ, AWQ, HQQ, GGUF | | Evaluation | evaluation@ai-research-skills | lm-eval-harness, BigCode, NeMo Evaluator | | Inference | inference-serving@ai-research-skills | vLLM, TensorRT-LLM, llama.cpp, SGLang | | MLOps | mlops@ai-research-skills | W&B, MLflow, TensorBoard | | Agents | agents@ai-research-skills | LangChain, LlamaIndex, CrewAI, AutoGPT | | RAG | rag@ai-research-skills | Chroma, FAISS, Pinecone, Qdrant, Sentence Transformers | | Prompt Eng | prompt-engineering@ai-research-skills | DSPy, Instructor, Guidance, Outlines | | Observability | observability@ai-research-skills | LangSmith, Phoenix | | Multimodal | multimodal@ai-research-skills | CLIP, Whisper, LLaVA, BLIP-2, SAM, Stable Diffusion, AudioCraft | | Emerging | emerging-techniques@ai-research-skills | MoE, Model Merging, Long Context, Speculative Decoding, Distillation, Pruning | | ML Paper Writing | ml-paper-writing@ai-research-skills | ML Paper Writing (LaTeX templates, citation verification, writing guides) |

🏗️ Model Architecture (5 skills)

  • LitGPT - Lightning AI's 20+ clean LLM implementations with production training recipes (462 lines + 4 refs)
  • Mamba - State-space models with O(n) complexity, 5× faster than Transformers (253 lines + 3 refs)
  • RWKV - RNN+Transformer hybrid, infinite context, Linux Foundation project (253 lines + 3 refs)
  • NanoGPT - Educational GPT in ~300 lines by Karpathy (283 lines + 3 refs)

🔤 Tokenization (2 skills)

  • HuggingFace Tokenizers - Rust-based, <20s/GB, BPE/WordPiece/Unigram algorithms (486 lines + 4 refs)
  • SentencePiece - Language-independent, 50k sentences/sec, used by T5/ALBERT (228 lines + 2 refs)

🎯 Fine-Tuning (4 skills)

  • Axolotl - YAML-based fine-tuning with 100+ models (156 lines + 4 refs)
  • LLaMA-Factory - WebUI no-code fine-tuning (78 lines + 5 refs)
  • Unsloth - 2x faster QLoRA fine-tuning (75 lines + 4 refs)
  • PEFT - Parameter-efficient fine-tuning with LoRA, QLoRA, DoRA, 25+ methods (431 lines + 2 refs)

🔬 Mechanistic Interpretability (4 skills)

  • TransformerLens - Neel Nanda's library for mech interp with HookPoints, activation caching (346 lines + 3 refs)
  • SAELens - Sparse Autoencoder training and analysis for feature discovery (386 lines + 3 refs)
  • pyvene - Stanford's causal intervention library with declarative configs (473 lines + 3 refs)
  • nnsight - Remote interpretability via NDIF, run experiments on 70B+ models (436 lines + 3 refs)

📊 Data Processing (2 skills)

  • Ray Data - Distributed ML data processing, streaming execution, GPU support (318 lines + 2 refs)
  • NeMo Curator - GPU-accelerated data curation, 16× faster deduplication (375 lines + 2 refs)

🎓 Post-Training (4 skills)

  • TRL Fine-Tuning - Transformer Reinforcement Learning (447 lines + 4 refs)
  • GRPO-RL-Training (TRL) - Group Relative Policy Optimization with TRL (569 lines, gold standard)
  • OpenRLHF - Full RLHF pipeline with Ray + vLLM (241 lines + 4 refs)
  • SimPO - Simple Preference Optimization, no reference model needed (211 lines + 3 refs)

🛡️ Safety & Alignment (3 skills)

⚡ Distributed Training (5 skills)

  • Megatron-Core - NVIDIA's framework for training 2B-462B param models with 47% MFU on H100 (359 lines + 4 refs)
  • DeepSpeed - Microsoft's ZeRO optimization (137 lines + 9 refs)
  • PyTorch FSDP - Fully Sharded Data Parallel (124 lines + 2 refs)
  • Accelerate - HuggingFace's 4-line distributed training API (324 lines + 3 refs)
  • PyTorch Lightning - High-level training framework with Trainer class (339 lines + 3 refs)
  • Ray Train - Multi-node orchestration and hyperparameter
五维分析
清晰度8/10
创新性8/10
实用性9/10
完整性8/10
可维护性7/10
优缺点分析

优点

  • 涵盖多个AI研究领域的全面技能。
  • 提供真实代码示例的专家级指导。
  • 支持AI研究实验的自主实施。

缺点

  • 对于初学者来说,安装过程可能比较复杂。
  • 由于技能数量庞大,文档可能会让人感到不知所措。
  • 对特定CLI工具的依赖可能限制可访问性。

相关技能

specweave

A
toolCo-Pilot / 辅助式
86/ 100

“就像有一个永不休息的私人助理,但确实需要Node.js才能运行。”

mgrep

A
toolCo-Pilot / 辅助式
86/ 100

“看起来很能打,但别让配置把人劝退。”

context-fundamentals

A
toolCo-Pilot / 辅助式
86/ 100

“这是你的智能体开始胡言乱语之前你希望拥有的教科书,但读了它也修不好那个bug。”

免责声明:本内容来源于 GitHub 开源项目,仅供展示和评分分析使用。

版权归原作者所有 zechenzhangAGI.