@@ -1420,7 +1420,7 @@ def _checkpoint_filter_fn(state_dict, model):
1420
1420
return state_dict # non-original checkpoint, no remapping needed
1421
1421
1422
1422
state_dict = state_dict .get ('state_dict' , state_dict )
1423
- if 'image_encoder.model.head.proj ' in state_dict :
1423
+ if 'image_encoder.model.patch_embed.0.rbr_conv.0.conv.weight ' in state_dict :
1424
1424
# remap MobileCLIP checkpoints
1425
1425
prefix = 'image_encoder.model.'
1426
1426
else :
@@ -1598,6 +1598,7 @@ def fastvit_mci0(pretrained=False, **kwargs):
1598
1598
se_downsamples = (False , False , True , True ),
1599
1599
pos_embs = (None , None , None , partial (RepConditionalPosEnc , spatial_shape = (7 , 7 ))),
1600
1600
token_mixers = ("repmixer" , "repmixer" , "repmixer" , "attention" ),
1601
+ lkc_use_act = True ,
1601
1602
)
1602
1603
return _create_fastvit ('fastvit_mci0' , pretrained = pretrained , ** dict (model_args , ** kwargs ))
1603
1604
@@ -1612,6 +1613,7 @@ def fastvit_mci1(pretrained=False, **kwargs):
1612
1613
se_downsamples = (False , False , True , True ),
1613
1614
pos_embs = (None , None , None , partial (RepConditionalPosEnc , spatial_shape = (7 , 7 ))),
1614
1615
token_mixers = ("repmixer" , "repmixer" , "repmixer" , "attention" ),
1616
+ lkc_use_act = True ,
1615
1617
)
1616
1618
return _create_fastvit ('fastvit_mci1' , pretrained = pretrained , ** dict (model_args , ** kwargs ))
1617
1619
@@ -1626,5 +1628,6 @@ def fastvit_mci2(pretrained=False, **kwargs):
1626
1628
se_downsamples = (False , False , True , True ),
1627
1629
pos_embs = (None , None , None , partial (RepConditionalPosEnc , spatial_shape = (7 , 7 ))),
1628
1630
token_mixers = ("repmixer" , "repmixer" , "repmixer" , "attention" ),
1631
+ lkc_use_act = True ,
1629
1632
)
1630
- return _create_fastvit ('fastvit_mci2' , pretrained = pretrained , ** dict (model_args , ** kwargs ))
1633
+ return _create_fastvit ('fastvit_mci2' , pretrained = pretrained , ** dict (model_args , ** kwargs ))
0 commit comments