Skip to content

[web] enable native EH for JSEP/Asyncify build #10392

[web] enable native EH for JSEP/Asyncify build

[web] enable native EH for JSEP/Asyncify build #10392

Triggered via pull request February 11, 2026 22:27
Status Failure
Total duration 37m 23s
Artifacts

windows_cuda.yml

on: pull_request
Windows GPU CUDA CI Pipeline
35m 0s
Windows GPU CUDA CI Pipeline
Windows GPU CUDA CI Pipeline Test Job
0s
Windows GPU CUDA CI Pipeline Test Job
Fit to window
Zoom out
Zoom in

Annotations

10 errors and 6 warnings
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L219
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const U *'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L218
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const U *'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L217
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const T *'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L216
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const T *'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L215
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const T *'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U> onnxruntime::contrib::cuda::GroupQueryAttentionData(onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U>)': expects 1 arguments - 0 provided
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U> onnxruntime::contrib::cuda::GroupQueryAttentionData(void)': could not deduce template argument for 'U'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U> onnxruntime::contrib::cuda::GroupQueryAttentionData(void)': could not deduce template argument for 'T'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
cannot deduce template arguments for 'onnxruntime::contrib::cuda::GroupQueryAttentionData'
Windows GPU CUDA CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData': too few template arguments
Windows GPU CUDA CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1234
epilog offset from end of function exceeds 4095
Windows GPU CUDA CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1227
epilog offset from end of function exceeds 4095
Windows GPU CUDA CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1220
epilog offset from end of function exceeds 4095
Windows GPU CUDA CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1213
epilog offset from end of function exceeds 4095
Windows GPU CUDA CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1206
epilog offset from end of function exceeds 4095
Windows GPU CUDA CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1199
epilog offset from end of function exceeds 4095