English | 简体中文
Pedestrian attribute recognition has been widely used in the intelligent community, industrial, and transportation monitoring. Many attribute recognition modules have been gathered in PP-Human, including gender, age, hats, eyes, clothing and up to 26 attributes in total. Also, the pre-trained models are offered here and users can download and use them directly.
Task | Algorithm | Precision | Inference Speed(ms) | Download Link |
---|---|---|---|---|
Pedestrian Detection/ Tracking | PP-YOLOE | mAP: 56.3 MOTA: 72.0 |
Detection: 28ms Tracking:33.1ms |
Download Link |
Pedestrian Attribute Analysis | StrongBaseline | ma: 94.86 | Per Person 2ms | Download Link |
- The precision of detection/ tracking models is obtained by training and testing on the dataset consist of MOT17, CrowdHuman, HIEVE, and some business data.
- The precision of pedestiran attribute analysis is obtained by training and testing on the dataset consist of PA100k, RAPv2, PETA, and some business data.
- The inference speed is T4, the speed of using TensorRT FP16.
- Download the model from the link in the above table, and unzip it to
./output_inference
. - When inputting the image, run the command as follows:
python deploy/pphuman/pipeline.py --config deploy/pphuman/config/infer_cfg.yml \
--image_file=test_image.jpg \
--device=gpu \
--enable_attr=True
- When inputting the video, run the command as follows:
python deploy/pphuman/pipeline.py --config deploy/pphuman/config/infer_cfg.yml \
--video_file=test_video.mp4 \
--device=gpu \
--enable_attr=True
-
If you want to change the model path, there are two methods:
- In
./deploy/pphuman/config/infer_cfg.yml
you can configurate different model paths. In attribute recognition models, you can modify the configuration in the field of ATTR. - Add
--model_dir
in the command line to change the model path:
- In
python deploy/pphuman/pipeline.py --config deploy/pphuman/config/infer_cfg.yml \
--video_file=test_video.mp4 \
--device=gpu \
--enable_attr=True \
--model_dir det=ppyoloe/
The test result is:
Data Source and Copyright:Skyinfor Technology. Thanks for the provision of actual scenario data, which are only used for academic research here.
- The PP-YOLOE model is used to handle detection boxs of input images/videos from object detection/ multi-object tracking. For details, please refer to the document PP-YOLOE.
- Capture every pedestrian in the input images with the help of coordiantes of detection boxes.
- Analyze the listed labels of pedestirans through attribute recognition. They are the same as those in the PA100k dataset. The label list is as follows:
- Gender
- Age: Less than 18; 18-60; Over 60
- Orientation: Front; Back; Side
- Accessories: Glasses; Hat; None
- HoldObjectsInFront: Yes; No
- Bag: BackPack; ShoulderBag; HandBag
- TopStyle: UpperStride; UpperLogo; UpperPlaid; UpperSplice
- BottomStyle: LowerStripe; LowerPattern
- ShortSleeve: Yes; No
- LongSleeve: Yes; No
- LongCoat: Yes; No
- Trousers: Yes; No
- Shorts: Yes; No
- Skirt&Dress: Yes; No
- Boots: Yes; No
- The model adopted in the attribute recognition is StrongBaseline, where the structure is the multi-class network structure based on ResNet50, and Weighted BCE loss and EMA are introduced for effect optimization.
@article{jia2020rethinking,
title={Rethinking of pedestrian attribute recognition: Realistic datasets with efficient method},
author={Jia, Jian and Huang, Houjing and Yang, Wenjie and Chen, Xiaotang and Huang, Kaiqi},
journal={arXiv preprint arXiv:2005.11909},
year={2020}
}