Skip to content

Commit 3fa1179

Browse files
committed
fix lint
1 parent fde29f4 commit 3fa1179

6 files changed

Lines changed: 5 additions & 6 deletions

File tree

python/sgl_jax/srt/configs/kimi_linear.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -142,4 +142,4 @@ def linear_layer_ids(self):
142142

143143
@property
144144
def full_attention_layer_ids(self):
145-
return [i for i in range(self.num_hidden_layers) if not self.is_kda_layer(i)]
145+
return [i for i in range(self.num_hidden_layers) if not self.is_kda_layer(i)]

python/sgl_jax/srt/hf_transformers_utils.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,6 @@
1919
from transformers.models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES
2020

2121
from sgl_jax.srt.configs.kimi_linear import KimiLinearConfig
22-
2322
from sgl_jax.srt.managers.tiktoken_tokenizer import TiktokenTokenizer
2423
from sgl_jax.srt.utils.common_utils import is_remote_url, lru_cache_frozenset
2524

python/sgl_jax/srt/models/kimi_linear.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -838,4 +838,4 @@ def _create_layer_mappings(
838838
return mappings
839839

840840

841-
EntryClass = KimiLinearForCausalLM
841+
EntryClass = KimiLinearForCausalLM

python/sgl_jax/srt/models/mimo_v2_flash.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -816,7 +816,7 @@ def __call__(
816816
else:
817817
output = self.logits_processor(hidden_states, self.model.embed_tokens, logits_metadata)
818818

819-
return output, {"token_to_kv_pool": layers_kv_fused},True, layers_topk_ids
819+
return output, {"token_to_kv_pool": layers_kv_fused}, True, layers_topk_ids
820820

821821

822822
EntryClass = [MiMoV2FlashForCausalLM]

python/sgl_jax/srt/models/qwen2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -500,7 +500,7 @@ def __call__(
500500
output = self.logits_processor(hidden_states, self.lm_head, logits_metadata)
501501
else:
502502
output = self.logits_processor(hidden_states, self.model.embed_tokens, logits_metadata)
503-
return output, {"token_to_kv_pool": layers_kv_fused},layers_callback_flag, None
503+
return output, {"token_to_kv_pool": layers_kv_fused}, layers_callback_flag, None
504504

505505

506506
EntryClass = Qwen2ForCausalLM

python/sgl_jax/srt/models/qwen3.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -589,7 +589,7 @@ def __call__(
589589
aux_hidden_states=aux_hidden_states,
590590
)
591591

592-
return output, {"token_to_kv_pool": layers_kv_fused},layers_callback_flag, None
592+
return output, {"token_to_kv_pool": layers_kv_fused}, layers_callback_flag, None
593593

594594

595595
EntryClass = Qwen3ForCausalLM

0 commit comments

Comments
 (0)