ai-research-skills
💡 摘要
该库提供了一整套全面的AI研究工程技能,用于构建和部署AI代理。
🎯 适合人群
🤖 AI 吐槽: “看起来很能打,但别让配置把人劝退。”
自述文件没有明确提到安全措施,这引发了对潜在风险的担忧,例如依赖漏洞和未经授权的访问。实施依赖管理工具和定期安全审计可以减轻这些风险。
AI Research Engineering Skills Library
The most comprehensive open-source library of AI research engineering skills for AI agents
77 Skills Powering AI Research in 2026
| | | | |:---:|:---:|:---:| | Model Architecture (5) | Fine-Tuning (4) | Post-Training (4) | | Distributed Training (5) | Optimization (6) | Inference (4) | | Tokenization (2) | Data Processing (2) | Evaluation (3) | | Safety & Alignment (3) | Agents (4) | RAG (5) | | Multimodal (7) | Prompt Engineering (4) | MLOps (3) | | Observability (2) | Infrastructure (3) | Mech Interp (4) | | Emerging Techniques (6) | ML Paper Writing (1) | |
Table of Contents
- Our Mission
- Path Towards AI Research Agent
- Available AI Research Engineering Skills
- Demo
- Skill Structure
- Roadmap
- Repository Structure
- Use Cases
Our Mission
We provide the layer of Engineering Ability that enable your coding agent to write and conduct AI research experiments, including preparing datasets, executing training pipelines, deploying models, and building your AI agents.
Path Towards AI Research Agent
Modern AI research requires mastering dozens of specialized tools and frameworks. AI Researchers spend more time debugging infrastructure than testing hypotheses—slowing the pace of scientific discovery. We provide a comprehensive library of expert-level research engineering skills that enable AI agents to autonomously implement and execute different stages of AI research experiments—from data preparation and model training to evaluation and deployment.
- Specialized Expertise - Each skill provides deep, production-ready knowledge of a specific framework (Megatron-LM, vLLM, TRL, etc.)
- End-to-End Coverage - 77 skills spanning model architecture, tokenization, fine-tuning, mechanistic interpretability, data processing, post-training, distributed training, optimization, evaluation, inference, infrastructure, agents, RAG, multimodal, prompt engineering, MLOps, observability, emerging techniques, and ML paper writing
- Research-Grade Quality - Documentation sourced from official repos, real GitHub issues, and battle-tested production workflows
Available AI Research Engineering Skills
Quality over quantity: Each skill provides comprehensive, expert-level guidance with real code examples, troubleshooting guides, and production-ready workflows.
📦 Install from Claude Code Marketplace
Install skill categories directly using the Claude Code CLI:
# Add the marketplace /plugin marketplace add zechenzhangAGI/AI-research-SKILLs # Install by category (20 categories available) /plugin install fine-tuning@ai-research-skills # Axolotl, LLaMA-Factory, PEFT, Unsloth /plugin install post-training@ai-research-skills # TRL, GRPO, OpenRLHF, SimPO /plugin install inference-serving@ai-research-skills # vLLM, TensorRT-LLM, llama.cpp, SGLang /plugin install distributed-training@ai-research-skills /plugin install optimization@ai-research-skills
All 20 Categories:
| Category | Install Command | Skills Included |
|----------|-----------------|-----------------|
| Model Architecture | model-architecture@ai-research-skills | LitGPT, Mamba, NanoGPT, RWKV |
| Tokenization | tokenization@ai-research-skills | HuggingFace Tokenizers, SentencePiece |
| Fine-Tuning | fine-tuning@ai-research-skills | Axolotl, LLaMA-Factory, PEFT, Unsloth |
| Mech Interp | mechanistic-interpretability@ai-research-skills | TransformerLens, SAELens, pyvene, nnsight |
| Data Processing | data-processing@ai-research-skills | NeMo Curator, Ray Data |
| Post-Training | post-training@ai-research-skills | TRL, GRPO, OpenRLHF, SimPO |
| Safety | safety-alignment@ai-research-skills | Constitutional AI, LlamaGuard, NeMo Guardrails |
| Distributed | distributed-training@ai-research-skills | DeepSpeed, FSDP, Accelerate, Megatron, Lightning, Ray Train |
| Infrastructure | infrastructure@ai-research-skills | Modal, Lambda Labs, SkyPilot |
| Optimization | optimization@ai-research-skills | Flash Attention, bitsandbytes, GPTQ, AWQ, HQQ, GGUF |
| Evaluation | evaluation@ai-research-skills | lm-eval-harness, BigCode, NeMo Evaluator |
| Inference | inference-serving@ai-research-skills | vLLM, TensorRT-LLM, llama.cpp, SGLang |
| MLOps | mlops@ai-research-skills | W&B, MLflow, TensorBoard |
| Agents | agents@ai-research-skills | LangChain, LlamaIndex, CrewAI, AutoGPT |
| RAG | rag@ai-research-skills | Chroma, FAISS, Pinecone, Qdrant, Sentence Transformers |
| Prompt Eng | prompt-engineering@ai-research-skills | DSPy, Instructor, Guidance, Outlines |
| Observability | observability@ai-research-skills | LangSmith, Phoenix |
| Multimodal | multimodal@ai-research-skills | CLIP, Whisper, LLaVA, BLIP-2, SAM, Stable Diffusion, AudioCraft |
| Emerging | emerging-techniques@ai-research-skills | MoE, Model Merging, Long Context, Speculative Decoding, Distillation, Pruning |
| ML Paper Writing | ml-paper-writing@ai-research-skills | ML Paper Writing (LaTeX templates, citation verification, writing guides) |
🏗️ Model Architecture (5 skills)
- LitGPT - Lightning AI's 20+ clean LLM implementations with production training recipes (462 lines + 4 refs)
- Mamba - State-space models with O(n) complexity, 5× faster than Transformers (253 lines + 3 refs)
- RWKV - RNN+Transformer hybrid, infinite context, Linux Foundation project (253 lines + 3 refs)
- NanoGPT - Educational GPT in ~300 lines by Karpathy (283 lines + 3 refs)
🔤 Tokenization (2 skills)
- HuggingFace Tokenizers - Rust-based, <20s/GB, BPE/WordPiece/Unigram algorithms (486 lines + 4 refs)
- SentencePiece - Language-independent, 50k sentences/sec, used by T5/ALBERT (228 lines + 2 refs)
🎯 Fine-Tuning (4 skills)
- Axolotl - YAML-based fine-tuning with 100+ models (156 lines + 4 refs)
- LLaMA-Factory - WebUI no-code fine-tuning (78 lines + 5 refs)
- Unsloth - 2x faster QLoRA fine-tuning (75 lines + 4 refs)
- PEFT - Parameter-efficient fine-tuning with LoRA, QLoRA, DoRA, 25+ methods (431 lines + 2 refs)
🔬 Mechanistic Interpretability (4 skills)
- TransformerLens - Neel Nanda's library for mech interp with HookPoints, activation caching (346 lines + 3 refs)
- SAELens - Sparse Autoencoder training and analysis for feature discovery (386 lines + 3 refs)
- pyvene - Stanford's causal intervention library with declarative configs (473 lines + 3 refs)
- nnsight - Remote interpretability via NDIF, run experiments on 70B+ models (436 lines + 3 refs)
📊 Data Processing (2 skills)
- Ray Data - Distributed ML data processing, streaming execution, GPU support (318 lines + 2 refs)
- NeMo Curator - GPU-accelerated data curation, 16× faster deduplication (375 lines + 2 refs)
🎓 Post-Training (4 skills)
- TRL Fine-Tuning - Transformer Reinforcement Learning (447 lines + 4 refs)
- GRPO-RL-Training (TRL) - Group Relative Policy Optimization with TRL (569 lines, gold standard)
- OpenRLHF - Full RLHF pipeline with Ray + vLLM (241 lines + 4 refs)
- SimPO - Simple Preference Optimization, no reference model needed (211 lines + 3 refs)
🛡️ Safety & Alignment (3 skills)
- Constitutional AI - AI-driven self-improvement via principles (282 lines)
- LlamaGuard - Safety classifier for LLM inputs/outputs (329 lines)
- NeMo Guardrails - Programmable guardrails with Colang (289 lines)
⚡ Distributed Training (5 skills)
- Megatron-Core - NVIDIA's framework for training 2B-462B param models with 47% MFU on H100 (359 lines + 4 refs)
- DeepSpeed - Microsoft's ZeRO optimization (137 lines + 9 refs)
- PyTorch FSDP - Fully Sharded Data Parallel (124 lines + 2 refs)
- Accelerate - HuggingFace's 4-line distributed training API (324 lines + 3 refs)
- PyTorch Lightning - High-level training framework with Trainer class (339 lines + 3 refs)
- Ray Train - Multi-node orchestration and hyperparameter
优点
- 涵盖多个AI研究领域的全面技能。
- 提供真实代码示例的专家级指导。
- 支持AI研究实验的自主实施。
缺点
- 对于初学者来说,安装过程可能比较复杂。
- 由于技能数量庞大,文档可能会让人感到不知所措。
- 对特定CLI工具的依赖可能限制可访问性。
相关技能
免责声明:本内容来源于 GitHub 开源项目,仅供展示和评分分析使用。
版权归原作者所有 zechenzhangAGI.
