@@ -2791,6 +2791,9 @@ def _cfg(url: str = '', **kwargs) -> Dict[str, Any]:
27912791 'vit_little_patch16_reg4_gap_256.sbb_in1k' : _cfg (
27922792 hf_hub_id = 'timm/' ,
27932793 input_size = (3 , 256 , 256 ), crop_pct = 0.95 ),
2794+ 'vit_dlittle_patch16_reg1_gap_256.sbb_nadamuon_in1k' : _cfg (
2795+ hf_hub_id = 'timm/' ,
2796+ input_size = (3 , 256 , 256 ), crop_pct = 0.95 ),
27942797 'vit_medium_patch16_reg1_gap_256.sbb_in1k' : _cfg (
27952798 hf_hub_id = 'timm/' ,
27962799 input_size = (3 , 256 , 256 ), crop_pct = 0.95 ),
@@ -4324,6 +4327,17 @@ def vit_little_patch16_reg1_gap_256(pretrained: bool = False, **kwargs) -> Visio
43244327 return model
43254328
43264329
4330+ @register_model
4331+ def vit_dlittle_patch16_reg1_gap_256 (pretrained : bool = False , ** kwargs ) -> VisionTransformer :
4332+ model_args = dict (
4333+ patch_size = 16 , embed_dim = 320 , depth = 14 , num_heads = 5 , init_values = 1e-5 , mlp_ratio = 5.6 ,
4334+ class_token = False , no_embed_class = True , reg_tokens = 1 , global_pool = 'avg' , attn_layer = 'diff' ,
4335+ )
4336+ model = _create_vision_transformer (
4337+ 'vit_dlittle_patch16_reg1_gap_256' , pretrained = pretrained , ** dict (model_args , ** kwargs ))
4338+ return model
4339+
4340+
43274341@register_model
43284342def vit_little_patch16_reg4_gap_256 (pretrained : bool = False , ** kwargs ) -> VisionTransformer :
43294343 model_args = dict (
0 commit comments