Skip to content

[CUDA] Support FP8 (E4M3) KV Cache for Group Query Attention #8715

[CUDA] Support FP8 (E4M3) KV Cache for Group Query Attention

[CUDA] Support FP8 (E4M3) KV Cache for Group Query Attention #8715

Annotations

3 warnings

6c. Build Extended Minimal (No Optional Features)

succeeded Feb 14, 2026 in 7m 30s