Skip to content

Soul-AILab/SoulX-FlashHead

Repository files navigation

SoulX-FlashHead: Oracle-guided Generation of Infinite Real-time Streaming Talking Heads

Tan Yu*, Qian Qiao*โœ‰, Le Shen*, Ke Zhou, Jincheng Hu, Dian Sheng, Bo Hu, Haoming Qin, Jun Gao, Changhai Zhou, Shunshun Yin, Siyuan Liu โœ‰

*Equal Contribution โœ‰Corresponding Author

Datasetย 

โšก Highlights

  • Model_Lite Released get 96 FPS, or 3-concurrent real-time(25+ FPS) streaming on single RTX4090.
  • Model_Pro Released can generate high-quality videos with 10.8 FPS on single RTX4090, or real-time(25+ FPS) on two RTX5090.
  • Model_Pretrained is coming soon, providing high-performance weights and experimental foundations for community research.

๐Ÿ”ฅ News

  • 2026.03.09 - Online demo on HuggingFace is available now. You can try it out directly.
  • 2026.03.04 - Gradio app is available now. Both common and streaming mode are supported.
  • 2026.03.02 - The ComfyUI node is now available. Thanks for the comfyui support of HM-RunningHub.
  • 2026.02.12 - The online demo is now available via the Soul App. Download it today to try it out.
  • 2026.02.12 - We have released the inference code, and the model weights.
  • 2026.02.12 - We released Project page on SoulX-FlashHead.
  • 2026.02.07 - We released Dataset.
  • 2026.02.07 - We released SoulX-FlashHead Technical Report on Arxiv and GitHub repository.

๐Ÿ“‘ Todo List

  • Technical report
  • Project Page
  • Inference code
  • Streaming online demo on HuggingFace
  • Distilled Checkpoint of Pro-Model & Lite-Model release
  • Pretrained Checkpoint release

๐ŸŒฐ Examples

More examples are available in the project.

qitiandasheng.mp4
chengdu.mp4
einstein.mp4

๐Ÿ“– Quickstart

๐Ÿ”ง Installation

1. Create a Conda environment

conda create -n flashhead python=3.10
conda activate flashhead

2. Install PyTorch on CUDA

pip install torch==2.7.1 torchvision==0.22.1 --index-url https://download.pytorch.org/whl/cu128

3. Install other dependencies

pip install -r requirements.txt

4. FlashAttention installation:

pip install ninja
pip install flash_attn==2.8.0.post2 --no-build-isolation

-- If it takes a long time, we recommend the way below.

  1. download wheel file from here
  2. pip install xxx.whl

5. SageAttention installation (Optional)

pip install sageattention==2.2.0 --no-build-isolation

6. FFmpeg installation

# Ubuntu / Debian
apt-get install ffmpeg
# CentOS / RHEL
yum install ffmpeg ffmpeg-devel

or

# Conda (no root required) 
conda install -c conda-forge ffmpeg==7

๐Ÿค— Model download

Model Component Description Link
SoulX-FlashHead-1_3B Our 1.3B model ๐Ÿค— Huggingface
wav2vec2-base-960h wav2vec2-base-960h ๐Ÿค— Huggingface
# If you are in china mainland, run this first: export HF_ENDPOINT=https://hf-mirror.com
pip install "huggingface_hub[cli]"
huggingface-cli download Soul-AILab/SoulX-FlashHead-1_3B --local-dir ./models/SoulX-FlashHead-1_3B
huggingface-cli download facebook/wav2vec2-base-960h --local-dir ./models/wav2vec2-base-960h

๐Ÿš€ Inference

# Infer with [Pro-Model] on single GPU
bash inference_script_single_gpu_pro.sh


# Infer with [Pro-Model] on multy GPUs
bash inference_script_multi_gpu_pro.sh
# Real-time inference speed of Pro-Model can only be supported on two RTX-5090 with SageAttention.


# Infer with [Lite-Model] on single GPU
bash inference_script_single_gpu_lite.sh
# Real-time inference speed can be supported on single RTX-4090 (up to 3 concurrent).

โšก๏ธ Gradio Demo

# Gradio support needs gradio==5.50.0, and Chrome recommonded.

# common gradio demo
python gradio_app.py

# streaming gradio demo (Only support single GPU)
python gradio_app_streaming.py

๐Ÿค— Streaming online demo

Click here to experience the real-time streaming demo on HuggingFace Spaces.

๐Ÿ‘‹ Online Experience

For a real-time interactive experience, scan the QR code to enter the event link. [2026.2.12~2026.3.11]

SoulApp event QR Code
Real-time Online Experience
(SoulApp ๅฎžๆ—ถๅœจ็บฟไฝ“้ชŒ)

๐Ÿ“ง Contact Us

If you are interested in leaving a message to our work, feel free to email [email protected] or [email protected] or [email protected] or [email protected] or [email protected]

We have opened a WeChat group. Additionally, we represent SoulApp and warmly welcome everyone to download the app and join our Soul group for further technical discussions and updates!

WeChat Group QR Code
Join WeChat Group
(ๅŠ ๅ…ฅๅพฎไฟกๆŠ€ๆœฏ็พค)
Soul App Group QR Code
Download SoulApp & Join Group
(ไธ‹่ฝฝSoulAppๅŠ ๅ…ฅ็พค็ป„)

๐Ÿ“š Citation

If you find our work useful in your research, please consider citing:

@misc{yu2026soulxflashheadoracleguidedgenerationinfinite,
      title={SoulX-FlashHead: Oracle-guided Generation of Infinite Real-time Streaming Talking Heads}, 
      author={Tan Yu and Qian Qiao and Le Shen and Ke Zhou and Jincheng Hu and Dian Sheng and Bo Hu and Haoming Qin and Jun Gao and Changhai Zhou and Shunshun Yin and Siyuan Liu},
      year={2026},
      eprint={2602.07449},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2602.07449}, 
}

๐Ÿ™‡ Acknowledgement

  • Wan: the base model we built upon.
  • LTX-Video: the VAE of our Lite-Model.
  • Self forcing: the codebase we built upon.
  • DMD and Self forcing++: the key distillation technique used by our method.
  • SoulX-FlashTalk is another model developed by our team, featuring 14B parameters and real-time capabilities.

Tip

If you find our work useful, please also consider starring the original repositories of these foundational methods.

๐Ÿ’ก Star History

Star History Chart

About

SoulX-FlashHead: A unified 1.3B-parameter framework designed for high-fidelity, infinite-length, and real-time streaming portrait video generation.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages