Extract human body keypoints from images or videos, construct temporal keypoint sequences, and provide structured data for downstream pose and motion analysis, recognition, and visualization.
PoseMotionAnalyzer is a lightweight project focused on human pose and motion analysis.
Rather than emphasizing single-frame pose detection, it focuses on:
- Structuring extracted body keypoints
- Organizing frame-level keypoints into temporal sequences
- Providing clean and reusable data interfaces for downstream analysis and modeling
The project is designed as an extensible engineering framework for pose and motion-related tasks.
The current implementation provides:
- Human body keypoint extraction
- Extract skeletal keypoints from images or video frames
- Keypoint sequence construction
- Assemble frame-level keypoints into temporal motion sequences
- Structured data output
- Unified data formats for analysis, modeling, and visualization
- Backend–frontend separation
- Backend for computation and data processing
- Frontend reserved for pose and motion visualization
PoseMotionAnalyzer/ ├─ backend/ │ └─ src/ │ ├─ config/ # Configuration files │ ├─ data/ # Input or intermediate data │ ├─ extractor/ # Keypoint and sequence extractors │ │ ├─ body_keypoint_extractor.py │ │ └─ body_keypoint_sequence_extractor.py │ ├─ models/ # Pose / motion models (reserved) │ ├─ utils/ # Utility functions │ └─ main.py # Main entry point ├─ test/ # Tests ├─ frontend/ │ ├─ app/ # Frontend application (reserved) │ └─ ui_design/ # UI designs ├─ .gitignore ├─ README.md ├─ requirements.txt └─ setup.ps1
- Input images or video data
- Extract human body keypoints using
body_keypoint_extractor - Construct temporal keypoint sequences with
body_keypoint_sequence_extractor - Use the outputs for:
- Motion analysis
- Pose comparison
- Visualization
- Downstream model training or inference
- Human pose analysis
- Motion sequence modeling
- Movement behavior research
- Pose and motion visualization
- Foundation module for action recognition systems
PoseMotionAnalyzer is a structured framework that converts visual human motion into analyzable temporal representations, emphasizing clarity, extensibility, and engineering practicality.