- π₯ Multimodal Video Captioning - Audio-Visual understanding
- ποΈ Computer Vision - 3D Reconstruction, Pose Estimation
- π€ Vision Transformers - Attention mechanisms for visual tasks
- π Deep Learning Research - PyTorch implementations
- π Portfolio | π§ ashokbk215@gmail.com
-
14:53
(UTC +05:45) - github.com/blazewild
- https://www.asokbk.com.np/
- in/asokbk
Highlights
- Pro
Pinned Loading
-
Real-Time-Motion-Transfer-to-a-3D-Avatar
Real-Time-Motion-Transfer-to-a-3D-Avatar PublicReal-time human pose detection and motion transfer to 3D avatars using MediaPipe, DNN, and Three.js β supports webcam and video inputs with custom avatar integration.
-
Custom_LLM_DataGen_Template
Custom_LLM_DataGen_Template Publicπ§ Modular pipeline for generating high-quality, domain-specific datasets for LLM fine-tuning β from PDFs and web scraping to synthetic Q&A generation, quality filtering, and training-ready formatting.
-
Hav-Cocap
Hav-Cocap PublicHav-Cocap: Hybrid Audio-Visual Compressed Video Captioning framework. Extends CoCap with an Audio Encoder and evaluated on the AVCaps dataset.
Jupyter Notebook
-
Blaze2Cap_AI_Motioner
Blaze2Cap_AI_Motioner Public3D Human Pose Estimation: BlazePose to TotalCapture Motion Dataset Pipeline with PyTorch DataLoader for motion capture research and machine learning
Python 2
-
GPT_FROM_SCRATCH
GPT_FROM_SCRATCH PublicMinimal GPT implementation from scratch using PyTorch β trains a character-level transformer on the Tiny Shakespeare dataset to demonstrate core LLM concepts.
Jupyter Notebook
-
MV_MAE
MV_MAE PublicMV-MAE is a hierarchical video model that leverages motion vectors and I-frames from compressed videos to efficiently learn masked motion representations for accurate UAV action recognition.
Python
If the problem persists, check the GitHub status page or contact support.


