Skip to content

Commit 940edf4

Browse files
committed
Add sbb vit 'dlittle' weights trained with NAdaMuon
1 parent d4ab516 commit 940edf4

File tree

1 file changed

+14
-0
lines changed

1 file changed

+14
-0
lines changed

timm/models/vision_transformer.py

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2791,6 +2791,9 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
27912791
'vit_little_patch16_reg4_gap_256.sbb_in1k': _cfg(
27922792
hf_hub_id='timm/',
27932793
input_size=(3, 256, 256), crop_pct=0.95),
2794+
'vit_dlittle_patch16_reg1_gap_256.sbb_nadamuon_in1k': _cfg(
2795+
hf_hub_id='timm/',
2796+
input_size=(3, 256, 256), crop_pct=0.95),
27942797
'vit_medium_patch16_reg1_gap_256.sbb_in1k': _cfg(
27952798
hf_hub_id='timm/',
27962799
input_size=(3, 256, 256), crop_pct=0.95),
@@ -4324,6 +4327,17 @@ def vit_little_patch16_reg1_gap_256(pretrained: bool = False, **kwargs) -> Visio
43244327
return model
43254328

43264329

4330+
@register_model
4331+
def vit_dlittle_patch16_reg1_gap_256(pretrained: bool = False, **kwargs) -> VisionTransformer:
4332+
model_args = dict(
4333+
patch_size=16, embed_dim=320, depth=14, num_heads=5, init_values=1e-5, mlp_ratio=5.6,
4334+
class_token=False, no_embed_class=True, reg_tokens=1, global_pool='avg', attn_layer='diff',
4335+
)
4336+
model = _create_vision_transformer(
4337+
'vit_dlittle_patch16_reg1_gap_256', pretrained=pretrained, **dict(model_args, **kwargs))
4338+
return model
4339+
4340+
43274341
@register_model
43284342
def vit_little_patch16_reg4_gap_256(pretrained: bool = False, **kwargs) -> VisionTransformer:
43294343
model_args = dict(

0 commit comments

Comments
 (0)