Skip to content

[web] enable native EH for JSEP/Asyncify build #10366

[web] enable native EH for JSEP/Asyncify build

[web] enable native EH for JSEP/Asyncify build #10366

Triggered via pull request February 11, 2026 22:27
Status Failure
Total duration 36m 8s
Artifacts

windows_tensorrt.yml

on: pull_request
Windows GPU TensorRT CI Pipeline
33m 45s
Windows GPU TensorRT CI Pipeline
Windows GPU TensorRT CI Pipeline Test Job
0s
Windows GPU TensorRT CI Pipeline Test Job
Fit to window
Zoom out
Zoom in

Annotations

10 errors and 6 warnings
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L219
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const U *'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L218
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const U *'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L217
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const T *'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L216
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const T *'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L215
'=': cannot convert from 'const onnxruntime::cuda::Attention<float>::ComputeInternal::CudaT *' to 'const T *'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U> onnxruntime::contrib::cuda::GroupQueryAttentionData(onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U>)': expects 1 arguments - 0 provided
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U> onnxruntime::contrib::cuda::GroupQueryAttentionData(void)': could not deduce template argument for 'U'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData<T,U> onnxruntime::contrib::cuda::GroupQueryAttentionData(void)': could not deduce template argument for 'T'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
cannot deduce template arguments for 'onnxruntime::contrib::cuda::GroupQueryAttentionData'
Windows GPU TensorRT CI Pipeline: onnxruntime/core/providers/cuda/llm/attention.cc#L199
'onnxruntime::contrib::cuda::GroupQueryAttentionData': too few template arguments
Windows GPU TensorRT CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1234
epilog offset from end of function exceeds 4095
Windows GPU TensorRT CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1227
epilog offset from end of function exceeds 4095
Windows GPU TensorRT CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1220
epilog offset from end of function exceeds 4095
Windows GPU TensorRT CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1213
epilog offset from end of function exceeds 4095
Windows GPU TensorRT CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1206
epilog offset from end of function exceeds 4095
Windows GPU TensorRT CI Pipeline: onnxruntime/core/mlas/lib/amd64/QgemmU8X8KernelAvx2.asm#L1199
epilog offset from end of function exceeds 4095