We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 3a89495 commit ae8f74bCopy full SHA for ae8f74b
1 file changed
tensorrt_llm/_torch/modules/attention.py
@@ -334,11 +334,11 @@ def __init__(
334
key="sparse_attention_config")
335
336
if config.sparse_attention_config.algorithm == "rocket":
337
- logger.warning("disable rope_fusion for RocketKV.")
+ logger.info_once("disable rope_fusion for RocketKV.")
338
self.rope_fusion = False
339
340
if self.rope_fusion and not attn_cls.support_fused_rope():
341
- logger.warning(
+ logger.info_once(
342
"rope_fusion is true but the attention backend does not support it. Will disable rope_fusion."
343
)
344
0 commit comments