|
| 1 | +# Awesome Human Activity Recognition |
| 2 | + |
| 3 | +> A curated, researcher-driven guide to **Human Activity Recognition** -- 53 datasets, key frameworks, pretrained models, tutorials, and benchmark tools across vision, wearable, skeleton, and multimodal modalities. |
| 4 | +
|
| 5 | +[](https://creativecommons.org/licenses/by/4.0/) |
| 6 | +[](https://github.com/Leo-Cyberautonomy/awesome-human-activity-recognition/pulls) |
| 7 | + |
| 8 | +## Quick Stats |
| 9 | + |
| 10 | +| Modality | Datasets | Highlights | |
| 11 | +|----------|----------|------------| |
| 12 | +| Vision (RGB/Depth) | 14 | Kinetics-700, UCF-101, ActivityNet, AVA | |
| 13 | +| Skeleton & MoCap | 7 | NTU RGB+D 60/120, AMASS, Human3.6M | |
| 14 | +| Wearable Sensors | 13 | UCI-HAR, PAMAP2, CAPTURE-24 (3883 hrs) | |
| 15 | +| Multimodal & Egocentric | 7 | Ego4D (3.3k hrs), EPIC-Kitchens-100 | |
| 16 | +| Emerging & Frontier | 12 | HumanML3D, Motion-X++, Ego-Exo4D | |
| 17 | + |
| 18 | +## Which Dataset Should I Use? |
| 19 | + |
| 20 | +!!! tip "Pick your modality and task, then follow the recommendation." |
| 21 | + |
| 22 | +=== "Video Classification" |
| 23 | + |
| 24 | + Start with **[Kinetics-700](../datasets/vision/kinetics-700.md)** for pretraining, evaluate on **[UCF-101](../datasets/vision/ucf101.md)** or **[HMDB-51](../datasets/vision/hmdb51.md)** for comparison with prior work. Browse all [Vision datasets](../datasets/vision/kinetics-700.md). |
| 25 | + |
| 26 | +=== "Temporal Action Detection" |
| 27 | + |
| 28 | + **[ActivityNet](../datasets/vision/activitynet.md)** for proposals, **[AVA](../datasets/vision/ava.md)** for spatio-temporal, **[MultiTHUMOS](../datasets/vision/multithumos.md)** for dense multi-label. |
| 29 | + |
| 30 | +=== "Skeleton / MoCap" |
| 31 | + |
| 32 | + **[NTU RGB+D 120](../datasets/vision/ntu-rgbd-120.md)** is the de facto standard. For text-motion alignment, use **[BABEL](../datasets/skeleton/babel.md)** or **[HumanML3D](../datasets/emerging/humanml3d.md)**. |
| 33 | + |
| 34 | +=== "Wearable Sensors" |
| 35 | + |
| 36 | + **[UCI-HAR](../datasets/wearable/uci-har.md)** for baselines, **[PAMAP2](../datasets/wearable/pamap2.md)** for multi-sensor, **[CAPTURE-24](../datasets/wearable/capture24.md)** for real-world scale (151 subjects, 3883 hours). |
| 37 | + |
| 38 | +=== "Egocentric / Multimodal" |
| 39 | + |
| 40 | + **[Ego4D](../datasets/multimodal/ego4d.md)** for scale (3.3k hours), **[EPIC-Kitchens-100](../datasets/multimodal/epic-kitchens-100.md)** for kitchen actions, **[Ego-Exo4D](../datasets/emerging/ego-exo4d.md)** for cross-view. |
| 41 | + |
| 42 | +=== "Text-to-Motion Generation" |
| 43 | + |
| 44 | + **[HumanML3D](../datasets/emerging/humanml3d.md)** for single-person, **[InterHuman](../datasets/emerging/interhuman.md)** for two-person, **[Motion-X++](../datasets/emerging/motionx-plus.md)** for whole-body with face and hands. |
| 45 | + |
| 46 | +## Explore |
| 47 | + |
| 48 | +- **[Datasets](../datasets/vision/kinetics-700.md)** -- Browse all 53 dataset cards organized by modality |
| 49 | +- **[Taxonomy](taxonomy.md)** -- Multi-dimensional classification of HAR approaches |
| 50 | +- **[Surveys](surveys.md)** -- Curated survey papers across all modalities |
| 51 | +- **[Benchmarking](benchmarking.md)** -- Compare datasets and methods |
| 52 | +- **[Roadmap](roadmap.md)** -- What is coming next |
| 53 | +- **[Contributing](../CONTRIBUTING.md)** -- How to add datasets or improve the list |
| 54 | + |
| 55 | +## Citation |
| 56 | + |
| 57 | +```bibtex |
| 58 | +@misc{awesome_har_2025, |
| 59 | + title = {Awesome Human Activity Recognition: A Curated List}, |
| 60 | + author = {Wenxuan Huang}, |
| 61 | + year = {2025}, |
| 62 | + url = {https://github.com/Leo-Cyberautonomy/awesome-human-activity-recognition}, |
| 63 | + note = {GitHub repository} |
| 64 | +} |
| 65 | +``` |
0 commit comments