Skip to content

Comprehensive open-source library of AI research and engineering skills for any AI model. Package the skills and your claude code/codex/gemini agent will be an AI research agent with full horsepower. Maintained by Orchestra Research.

License

Notifications You must be signed in to change notification settings

zechenzhangAGI/AI-research-SKILLs

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

58 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

AI Research Engineering Skills Library

The most comprehensive open-source library of AI research engineering skills for AI agents

License: MIT Blog Post Demo

Table of Contents

Our Mission

We provide the layer of Engineering Ability that enable your coding agent to write conduct AI research experiments, including preparing datasets, executing training pipelines, deploying models, and validating scientific hypotheses.

AI Research Agent System
System diagram of an AI research agent

Path Towards AI Research Agent

Modern AI research requires mastering dozens of specialized tools and frameworks. AI Researchers spend more time debugging infrastructure than testing hypothesesβ€”slowing the pace of scientific discovery. We provide a comprehensive library of expert-level research engineering skills that enable AI agents to autonomously implement and execute different stages of AI research experimentsβ€”from data preparation and model training to evaluation and deployment.

  • Specialized Expertise - Each skill provides deep, production-ready knowledge of a specific framework (Megatron-LM, vLLM, TRL, etc.)
  • End-to-End Coverage - 74 skills spanning model architecture, tokenization, fine-tuning, mechanistic interpretability, data processing, post-training, distributed training, optimization, inference, infrastructure, agents, RAG, multimodal, prompt engineering, MLOps, observability, and emerging techniques
  • Research-Grade Quality - Documentation sourced from official repos, real GitHub issues, and battle-tested production workflows

Available AI Research Engineering Skills

Quality over quantity: Each skill provides comprehensive, expert-level guidance with real code examples, troubleshooting guides, and production-ready workflows.

πŸ“¦ Install from Marketplace

Install individual skills directly from the marketplace using the Claude Code CLI:

# Install a single skill
/plugin install skill-name@ai-research-skills

# Examples:
/plugin install serving-llms-vllm@ai-research-skills
/plugin install grpo-rl-training@ai-research-skills
/plugin install langchain@ai-research-skills

πŸ—οΈ Model Architecture (5 skills)

  • LitGPT - Lightning AI's 20+ clean LLM implementations with production training recipes (462 lines + 4 refs)
  • Mamba - State-space models with O(n) complexity, 5Γ— faster than Transformers (253 lines + 3 refs)
  • RWKV - RNN+Transformer hybrid, infinite context, Linux Foundation project (253 lines + 3 refs)
  • NanoGPT - Educational GPT in ~300 lines by Karpathy (283 lines + 3 refs)

πŸ”€ Tokenization (2 skills)

  • HuggingFace Tokenizers - Rust-based, <20s/GB, BPE/WordPiece/Unigram algorithms (486 lines + 4 refs)
  • SentencePiece - Language-independent, 50k sentences/sec, used by T5/ALBERT (228 lines + 2 refs)

🎯 Fine-Tuning (4 skills)

  • Axolotl - YAML-based fine-tuning with 100+ models (156 lines + 4 refs)
  • LLaMA-Factory - WebUI no-code fine-tuning (78 lines + 5 refs)
  • Unsloth - 2x faster QLoRA fine-tuning (75 lines + 4 refs)
  • PEFT - Parameter-efficient fine-tuning with LoRA, QLoRA, DoRA, 25+ methods (431 lines + 2 refs)

πŸ”¬ Mechanistic Interpretability (4 skills)

  • TransformerLens - Neel Nanda's library for mech interp with HookPoints, activation caching (346 lines + 3 refs)
  • SAELens - Sparse Autoencoder training and analysis for feature discovery (386 lines + 3 refs)
  • pyvene - Stanford's causal intervention library with declarative configs (473 lines + 3 refs)
  • nnsight - Remote interpretability via NDIF, run experiments on 70B+ models (436 lines + 3 refs)

πŸ“Š Data Processing (2 skills)

  • Ray Data - Distributed ML data processing, streaming execution, GPU support (318 lines + 2 refs)
  • NeMo Curator - GPU-accelerated data curation, 16Γ— faster deduplication (375 lines + 2 refs)

πŸŽ“ Post-Training (4 skills)

  • TRL Fine-Tuning - Transformer Reinforcement Learning (447 lines + 4 refs)
  • GRPO-RL-Training (TRL) - Group Relative Policy Optimization with TRL (569 lines, gold standard)
  • OpenRLHF - Full RLHF pipeline with Ray + vLLM (241 lines + 4 refs)
  • SimPO - Simple Preference Optimization, no reference model needed (211 lines + 3 refs)

πŸ›‘οΈ Safety & Alignment (3 skills)

⚑ Distributed Training (5 skills)

  • Megatron-Core - NVIDIA's framework for training 2B-462B param models with 47% MFU on H100 (359 lines + 4 refs)
  • DeepSpeed - Microsoft's ZeRO optimization (137 lines + 9 refs)
  • PyTorch FSDP - Fully Sharded Data Parallel (124 lines + 2 refs)
  • Accelerate - HuggingFace's 4-line distributed training API (324 lines + 3 refs)
  • PyTorch Lightning - High-level training framework with Trainer class (339 lines + 3 refs)
  • Ray Train - Multi-node orchestration and hyperparameter tuning (399 lines + 1 ref)

πŸš€ Optimization (6 skills)

  • Flash Attention - 2-4x faster attention with memory efficiency (359 lines + 2 refs)
  • bitsandbytes - 8-bit/4-bit quantization for 50-75% memory reduction (403 lines + 3 refs)
  • GPTQ - 4-bit post-training quantization, 4Γ— memory reduction, <2% accuracy loss (443 lines + 3 refs)
  • AWQ - Activation-aware weight quantization, 4-bit with minimal accuracy loss (310 lines + 2 refs)
  • HQQ - Half-Quadratic Quantization, no calibration data needed, multi-backend (370 lines + 2 refs)
  • GGUF - llama.cpp quantization format, K-quant methods, CPU/Metal inference (380 lines + 2 refs)

πŸ“Š Evaluation (1 skill)

  • lm-evaluation-harness - EleutherAI's standard for benchmarking LLMs across 60+ tasks (482 lines + 4 refs)

☁️ Infrastructure (3 skills)

  • Modal - Serverless GPU cloud with Python-native API, T4-H200 on-demand (342 lines + 2 refs)
  • SkyPilot - Multi-cloud orchestration across 20+ providers with spot recovery (390 lines + 2 refs)
  • Lambda Labs - Reserved/on-demand GPU cloud with H100/A100, persistent filesystems (390 lines + 2 refs)

πŸ”₯ Inference & Serving (4 skills)

  • vLLM - High-throughput LLM serving with PagedAttention (356 lines + 4 refs, production-ready)
  • TensorRT-LLM - NVIDIA's fastest inference, 24k tok/s, FP8/INT4 quantization (180 lines + 3 refs)
  • llama.cpp - CPU/Apple Silicon inference, GGUF quantization (251 lines + 3 refs)
  • SGLang - Structured generation with RadixAttention, 5-10Γ— faster for agents (435 lines + 3 refs)

πŸ€– Agents (4 skills)

  • LangChain - Most popular agent framework, 500+ integrations, ReAct pattern (658 lines + 3 refs, production-ready)
  • LlamaIndex - Data framework for LLM apps, 300+ connectors, RAG-focused (535 lines + 3 refs)
  • CrewAI - Multi-agent orchestration, role-based collaboration, autonomous workflows (498 lines + 3 refs)
  • AutoGPT - Autonomous AI agent platform, visual workflow builder, continuous execution (400 lines + 2 refs)

πŸ” RAG (5 skills)

  • Chroma - Open-source embedding database, local/cloud, 24k stars (385 lines + 1 ref)
  • FAISS - Facebook's similarity search, billion-scale, GPU acceleration (295 lines)
  • Sentence Transformers - 5000+ embedding models, multilingual, 15k stars (370 lines)
  • Pinecone - Managed vector database, auto-scaling, <100ms latency (410 lines)
  • Qdrant - High-performance vector search, Rust-powered, hybrid search with filtering (493 lines + 2 refs)

🎨 Multimodal (7 skills)

  • CLIP - OpenAI's vision-language model, zero-shot classification, 25k stars (320 lines)
  • Whisper - Robust speech recognition, 99 languages, 73k stars (395 lines)
  • LLaVA - Vision-language assistant, image chat, GPT-4V level (360 lines)
  • Stable Diffusion - Text-to-image generation via HuggingFace Diffusers, SDXL, ControlNet (380 lines + 2 refs)
  • Segment Anything - Meta's SAM for zero-shot image segmentation with points/boxes (500 lines + 2 refs)
  • BLIP-2 - Vision-language pretraining with Q-Former, image captioning, VQA (500 lines + 2 refs)
  • AudioCraft - Meta's MusicGen/AudioGen for text-to-music and text-to-sound (470 lines + 2 refs)

🎯 Prompt Engineering (4 skills)

  • DSPy - Declarative prompt programming with optimizers, Stanford NLP, 22k stars (438 lines + 3 refs)
  • Instructor - Structured LLM outputs with Pydantic validation, 15k stars (726 lines + 3 refs)
  • Guidance - Constrained generation with regex/grammars, Microsoft Research, 18k stars (485 lines + 3 refs)
  • Outlines - Structured text with FSM, zero-overhead, 8k stars (601 lines + 3 refs)

πŸ“Š MLOps (3 skills)

  • Weights & Biases - Experiment tracking, sweeps, artifacts, model registry (427 lines + 3 refs)
  • MLflow - Model registry, tracking, deployment, autologging (514 lines + 3 refs)
  • TensorBoard - Visualization, profiling, embeddings, scalars/images (538 lines + 3 refs)

πŸ‘οΈ Observability (2 skills)

  • LangSmith - LLM observability, tracing, evaluation, monitoring for AI apps (422 lines + 2 refs)
  • Phoenix - Open-source AI observability with OpenTelemetry tracing and LLM evaluation (380 lines + 2 refs)

πŸ”¬ Emerging Techniques (6 skills)

  • MoE Training - Mixture of Experts training with DeepSpeed, Mixtral 8x7B, 5Γ— cost reduction (515 lines + 3 refs)
  • Model Merging - Combine models with TIES, DARE, SLERP using mergekit (528 lines + 3 refs)
  • Long Context - Extend context windows with RoPE, YaRN, ALiBi, 32k-128k tokens (624 lines + 3 refs)
  • Speculative Decoding - 1.5-3.6Γ— faster inference with Medusa, Lookahead (379 lines)
  • Knowledge Distillation - Compress models 70Bβ†’7B with MiniLLM, temperature scaling (424 lines)
  • Model Pruning - 50% sparsity with Wanda, SparseGPT, <1% accuracy loss (417 lines)

Available skills in Claude marketplace (74 total):

Category Skills
Model Architecture implementing-llms-litgpt, mamba-architecture, nanogpt, rwkv-architecture
Tokenization huggingface-tokenizers, sentencepiece
Fine-Tuning axolotl, llama-factory, peft-fine-tuning, unsloth
Mechanistic Interpretability transformer-lens-interpretability, sparse-autoencoder-training, pyvene-interventions, nnsight-remote-interpretability
Data Processing nemo-curator, ray-data
Post-Training grpo-rl-training, openrlhf-training, simpo-training, fine-tuning-with-trl
Safety constitutional-ai, llamaguard, nemo-guardrails
Distributed Training huggingface-accelerate, deepspeed, training-llms-megatron, pytorch-fsdp, pytorch-lightning, ray-train
Infrastructure lambda-labs-gpu-cloud, modal-serverless-gpu, skypilot-multi-cloud-orchestration
Optimization awq-quantization, quantizing-models-bitsandbytes, optimizing-attention-flash, gguf-quantization, gptq, hqq-quantization
Evaluation evaluating-llms-harness
Inference llama-cpp, sglang, tensorrt-llm, serving-llms-vllm
MLOps mlflow, tensorboard, weights-and-biases
Agents autogpt-agents, crewai-multi-agent, langchain, llamaindex
RAG chroma, faiss, pinecone, qdrant-vector-search, sentence-transformers
Prompt Engineering dspy, guidance, instructor, outlines
Observability langsmith-observability, phoenix-observability
Multimodal audiocraft-audio-generation, blip-2-vision-language, clip, llava, segment-anything-model, stable-diffusion-image-generation, whisper
Emerging Techniques knowledge-distillation, long-context, model-merging, model-pruning, moe-training, speculative-decoding

Demo

All 74 skills in this repo are automatically synced to Orchestra Research, where you can add them to your projects with one click and use them with AI research agents.

Demo: With this skills, a physics PhD is able to reproduce Thinking Machines Lab's "LoRA Without Regret" findings. The Orchestra agent autonomously wrote training code using TRL, provisioned H100 GPUs, ran GRPO experiments overnight, and generated publication-ready analysis, successfully validating that rank=16 LoRA achieves 99.4% of rank=256's SFT performance and that rank=1 LoRA outperforms full fine-tuning on RL tasks (52.1% vs 33.3% on GSM8k math reasoning). (Video demo)

Note: When you contribute a skill to this repo via PR, it automatically syncs to the Orchestra marketplace after merge.

πŸ› οΈ Alternative Usage Methods

For Claude Users (Claude.ai, Claude Code):

# 1. Download skill folder
cd 01-model-architecture/litgpt

# 2. Use directly in Claude Code workspace
# OR zip and upload to Claude.ai Projects
zip -r litgpt-skill.zip SKILL.md references/

For Other AI Coding Assistants:

  • Gemini CLI: Point to skill directory in your workspace
  • Grok Code: Use skill folder as context
  • Cursor/Windsurf: Add skill folder to project knowledge

For Custom RAG/Agent Systems:

  • Ingest SKILL.md + references/ into your knowledge base
  • Use as retrieval context for specialized queries
  • Build domain-specific agents with curated skill subsets

πŸ‘¨β€πŸ’» For Skill Creators

Your contributions power the entire ecosystem! When you contribute a skill to this repo:

  1. It automatically syncs to Orchestra marketplace
  2. Thousands of researchers can use your expertise
  3. AI agents become more capable at conducting research

Getting started:

  1. Read CONTRIBUTING.md - Step-by-step guide
  2. Use SKILL_TEMPLATE.md - Copy-paste scaffold
  3. Run validation: python scripts/validate_skill.py your-skill/
  4. Submit PR - We review within 48 hours, auto-publish to Orchestra on merge

Skill Structure

Each skill follows a battle-tested format for maximum usefulness:

skill-name/
β”œβ”€β”€ SKILL.md                    # Quick reference (50-150 lines)
β”‚   β”œβ”€β”€ Metadata (name, description, version)
β”‚   β”œβ”€β”€ When to use this skill
β”‚   β”œβ”€β”€ Quick patterns & examples
β”‚   └── Links to references
β”‚
β”œβ”€β”€ references/                 # Deep documentation (300KB+)
β”‚   β”œβ”€β”€ README.md              # From GitHub/official docs
β”‚   β”œβ”€β”€ api.md                 # API reference
β”‚   β”œβ”€β”€ tutorials.md           # Step-by-step guides
β”‚   β”œβ”€β”€ issues.md              # Real GitHub issues & solutions
β”‚   β”œβ”€β”€ releases.md            # Version history & breaking changes
β”‚   └── file_structure.md      # Codebase navigation
β”‚
β”œβ”€β”€ scripts/                    # Helper scripts (optional)
└── assets/                     # Templates & examples (optional)
Quality Standards
  • 300KB+ documentation from official sources
  • Real GitHub issues & solutions (when available)
  • Code examples with language detection
  • Version history & breaking changes
  • Links to official docs

Roadmap

We're building towards 80 comprehensive skills across the full AI research lifecycle. See our detailed roadmap for the complete development plan.

View Full Roadmap β†’

View Detailed Statistics
Metric Current Target
Skills 74 (high-quality, standardized YAML) 80
Avg Lines/Skill 420 lines (focused + progressive disclosure) 200-600 lines
Documentation ~120,000 lines total (SKILL.md + references) 100,000+ lines
Gold Standard Skills 62 with comprehensive references 50+
Contributors 1 100+
Coverage Architecture, Tokenization, Fine-Tuning, Mechanistic Interpretability, Data Processing, Post-Training, Safety, Distributed, Optimization, Evaluation, Infrastructure, Inference, Agents, RAG, Multimodal, Prompt Engineering, MLOps, Observability Full Lifecycle βœ…

Recent Progress: +4 skills (TransformerLens, SAELens, pyvene, nnsight) adding Mechanistic Interpretability category for reverse-engineering neural networks

Philosophy: Quality > Quantity. Following Anthropic official best practices - each skill provides 200-500 lines of focused, actionable guidance with progressive disclosure.

Repository Structure

claude-ai-research-skills/
β”œβ”€β”€ README.md                    ← You are here
β”œβ”€β”€ CONTRIBUTING.md              ← Contribution guide
β”œβ”€β”€ SKILL_TEMPLATE.md            ← Skill scaffold
β”œβ”€β”€ ROADMAP.md                   ← Development roadmap
β”‚
β”œβ”€β”€ 01-model-architecture/       (5 skills βœ“ - Megatron, LitGPT, Mamba, RWKV, NanoGPT)
β”œβ”€β”€ 02-tokenization/             (2 skills βœ“ - HuggingFace Tokenizers, SentencePiece)
β”œβ”€β”€ 03-fine-tuning/              (4 skills βœ“ - Axolotl, LLaMA-Factory, Unsloth, PEFT)
β”œβ”€β”€ 04-mechanistic-interpretability/ (4 skills βœ“ - TransformerLens, SAELens, pyvene, nnsight)
β”œβ”€β”€ 05-data-processing/          (2 skills βœ“ - Ray Data, NeMo Curator)
β”œβ”€β”€ 06-post-training/            (4 skills βœ“ - TRL, GRPO, OpenRLHF, SimPO)
β”œβ”€β”€ 07-safety-alignment/         (3 skills βœ“ - Constitutional AI, LlamaGuard, NeMo Guardrails)
β”œβ”€β”€ 08-distributed-training/     (5 skills βœ“ - DeepSpeed, FSDP, Accelerate, Lightning, Ray Train)
β”œβ”€β”€ 09-infrastructure/           (3 skills βœ“ - Modal, SkyPilot, Lambda Labs)
β”œβ”€β”€ 10-optimization/             (6 skills βœ“ - Flash Attention, bitsandbytes, GPTQ, AWQ, HQQ, GGUF)
β”œβ”€β”€ 11-evaluation/               (1 skill βœ“ - lm-evaluation-harness)
β”œβ”€β”€ 12-inference-serving/        (4 skills βœ“ - vLLM, TensorRT-LLM, llama.cpp, SGLang)
β”œβ”€β”€ 13-mlops/                    (3 skills βœ“ - Weights & Biases, MLflow, TensorBoard)
β”œβ”€β”€ 14-agents/                   (4 skills βœ“ - LangChain, LlamaIndex, CrewAI, AutoGPT)
β”œβ”€β”€ 15-rag/                      (5 skills βœ“ - Chroma, FAISS, Sentence Transformers, Pinecone, Qdrant)
β”œβ”€β”€ 16-prompt-engineering/       (4 skills βœ“ - DSPy, Instructor, Guidance, Outlines)
β”œβ”€β”€ 17-observability/            (2 skills βœ“ - LangSmith, Phoenix)
β”œβ”€β”€ 18-multimodal/               (7 skills βœ“ - CLIP, Whisper, LLaVA, Stable Diffusion, SAM, BLIP-2, AudioCraft)
└── 19-emerging-techniques/      (6 skills βœ“ - MoE, Model Merging, Long Context, Speculative Decoding, Distillation, Pruning)

Use Cases

For Researchers

"I need to fine-tune Llama 3 with custom data" β†’ 03-fine-tuning/axolotl/ - YAML configs, 100+ model support

For ML Engineers

"How do I optimize inference latency?" β†’ 12-inference-serving/vllm/ - PagedAttention, batching

For Students

"I want to learn how transformers work" β†’ 01-model-architecture/litgpt/ - Clean implementations

For Teams

"We need to scale training to 100 GPUs" β†’ 08-distributed-training/deepspeed/ - ZeRO stages, 3D parallelism

License

MIT License - See LICENSE for details.

Note: Individual skills may reference libraries with different licenses. Please check each project's license before use.

Acknowledgments

Built with:

  • Claude Code - AI pair programming
  • Skill Seeker - Automated doc scraping
  • Open Source AI Community - For amazing tools and docs

Special thanks to:

  • EleutherAI, HuggingFace, NVIDIA, Lightning AI, Meta AI, Anthropic
  • All researchers who maintain excellent documentation

Contributing

We welcome contributions from the AI research community! See CONTRIBUTING.md for detailed guidelines on:

  • Adding new skills
  • Improving existing skills
  • Quality standards and best practices
  • Submission process

All contributors are featured in our Contributors Hall of Fame 🌟

Recent Updates

December 2025 - v0.11.0 πŸ”¬ Mechanistic Interpretability
  • πŸ”¬ NEW CATEGORY: Mechanistic Interpretability (4 skills)
  • πŸ” TransformerLens skill: Neel Nanda's library for mech interp with HookPoints, activation caching, circuit analysis
  • 🧠 SAELens skill: Sparse Autoencoder training and analysis for feature discovery, monosemanticity research
  • ⚑ pyvene skill: Stanford's causal intervention library with declarative configs, DAS, activation patching
  • 🌐 nnsight skill: Remote interpretability via NDIF, run experiments on 70B+ models without local GPUs
  • πŸ“ ~6,500 new lines of documentation across 16 files
  • 74 total skills (filling the missing 04 category slot)
November 25, 2025 - v0.10.0 πŸŽ‰ 70 Skills Complete!
  • πŸŽ‰ ROADMAP COMPLETE: Reached 70-skill milestone!
  • πŸš€ Added 4 skills: Lambda Labs, Segment Anything (SAM), BLIP-2, AudioCraft
  • ☁️ Lambda Labs skill: Reserved/on-demand GPU cloud with H100/A100, persistent filesystems, 1-Click Clusters
  • πŸ–ΌοΈ SAM skill: Meta's Segment Anything for zero-shot image segmentation with points/boxes/masks
  • πŸ‘οΈ BLIP-2 skill: Vision-language pretraining with Q-Former, image captioning, VQA
  • 🎡 AudioCraft skill: Meta's MusicGen/AudioGen for text-to-music and text-to-sound generation
  • πŸ“ ~10,000 new lines of documentation across 12 files
  • 70 total skills (100% roadmap complete!)
November 25, 2025 - v0.9.0
  • πŸš€ Added 2 infrastructure skills: Modal, SkyPilot
  • ☁️ Modal skill: Serverless GPU cloud with Python-native API, T4-H200 on-demand, auto-scaling
  • 🌐 SkyPilot skill: Multi-cloud orchestration across 20+ providers with spot recovery
  • ✨ New Infrastructure category (2 skills - serverless GPU and multi-cloud orchestration)
  • πŸ“ ~2,500 new lines of documentation across 6 files
  • 66 total skills (94% towards 70-skill target)
November 25, 2025 - v0.8.0
  • πŸš€ Added 5 high-priority skills: HQQ, GGUF, Phoenix, AutoGPT, Stable Diffusion
  • ⚑ HQQ skill: Half-Quadratic Quantization without calibration data, multi-backend support
  • πŸ“¦ GGUF skill: llama.cpp quantization format, K-quant methods, CPU/Metal inference
  • πŸ‘οΈ Phoenix skill: Open-source AI observability with OpenTelemetry tracing and LLM evaluation
  • πŸ€– AutoGPT skill: Autonomous AI agent platform with visual workflow builder
  • 🎨 Stable Diffusion skill: Text-to-image generation via Diffusers, SDXL, ControlNet, LoRA
  • πŸ“ ~9,000 new lines of documentation across 15 files
  • 64 total skills (91% towards 70-skill target)
November 25, 2025 - v0.7.0
  • πŸš€ Added 5 high-priority skills: PEFT, CrewAI, Qdrant, AWQ, LangSmith
  • ✨ New Observability category with LangSmith for LLM tracing and evaluation
  • 🎯 PEFT skill: Parameter-efficient fine-tuning with LoRA, QLoRA, DoRA, 25+ methods
  • πŸ€– CrewAI skill: Multi-agent orchestration with role-based collaboration
  • πŸ” Qdrant skill: High-performance Rust vector search with hybrid filtering
  • ⚑ AWQ skill: Activation-aware 4-bit quantization with minimal accuracy loss
  • πŸ“ ~8,000 new lines of documentation across 15 files
  • 59 total skills (84% towards 70-skill target)
November 15, 2025 - v0.6.0
  • πŸ“Š Added 3 comprehensive MLOps skills: Weights & Biases, MLflow, TensorBoard
  • ✨ New MLOps category (3 skills - experiment tracking, model registry, visualization)
  • πŸ“ ~10,000 new lines of documentation across 13 files
  • πŸ”§ Comprehensive coverage: experiment tracking, hyperparameter sweeps, model registry, profiling, embeddings visualization
  • 54 total skills (77% towards 70-skill target)
November 12, 2025 - v0.5.0
  • 🎯 Added 4 comprehensive prompt engineering skills: DSPy, Instructor, Guidance, Outlines
  • ✨ New Prompt Engineering category (4 skills - DSPy, Instructor, Guidance, Outlines)
  • πŸ“ ~10,000 new lines of documentation across 16 files
  • πŸ”§ Comprehensive coverage: declarative programming, structured outputs, constrained generation, FSM-based generation
  • 47 total skills (67% towards 70-skill target)
November 9, 2025 - v0.4.0
  • πŸ€– Added 11 comprehensive skills: LangChain, LlamaIndex, Chroma, FAISS, Sentence Transformers, Pinecone, CLIP, Whisper, LLaVA
  • ✨ New Agents category (2 skills - LangChain, LlamaIndex)
  • πŸ” New RAG category (4 skills - Chroma, FAISS, Sentence Transformers, Pinecone)
  • 🎨 New Multimodal category (3 skills - CLIP, Whisper, LLaVA)
  • πŸ“ ~15,000 new lines of documentation
  • 43 total skills (61% towards 70-skill target)
November 8, 2025 - v0.3.0
  • πŸš€ Added 8 comprehensive skills: TensorRT-LLM, llama.cpp, SGLang, GPTQ, HuggingFace Tokenizers, SentencePiece, Ray Data, NeMo Curator
  • ⚑ Completed Inference & Serving category (4/4 skills)
  • πŸ”€ New Tokenization category (2 skills)
  • πŸ“Š New Data Processing category (2 skills)
  • πŸ“ 9,617 new lines of documentation across 30 files
  • 32 total skills (45% towards 70-skill target)
November 6, 2025 - v0.2.0
  • Added 10 skills from GitHub (Megatron-Core, Lightning, Ray Train, etc.)
  • Improved skill structure with comprehensive references
  • Created strategic roadmap to 70 skills
  • Added contribution guidelines
November 3, 2025 - v0.1.0
  • πŸŽ‰ Initial release with 5 fine-tuning skills

Star History

Star History Chart

About

Comprehensive open-source library of AI research and engineering skills for any AI model. Package the skills and your claude code/codex/gemini agent will be an AI research agent with full horsepower. Maintained by Orchestra Research.

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Packages

No packages published

Contributors 3

  •  
  •  
  •  

Languages