You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am currently trying to fine-tune an MAE based on pretrained VIT from timm. However, when I do:
v = timm.create_model('vit_base_patch16_224', pretrained=True)
num_ftrs = v.head.in_features
v.head = nn.Linear(num_ftrs, 2)
model = MAE(
encoder = v,
masking_ratio = 0.75, # the paper recommended 75% masked patches
decoder_dim = 512, # paper showed good results with just 512
decoder_depth = 6 # anywhere from 1 to 8
)
I got "AttributeError: 'VisionTransformer' object has no attribute 'pos_embedding'"
It seems that timm model is not compatible with the MAE implementation. Can this be easily fixed or I will have to change the internal implementation of MAE?
The text was updated successfully, but these errors were encountered:
Hi There,
I am currently trying to fine-tune an MAE based on pretrained VIT from timm. However, when I do:
I got "AttributeError: 'VisionTransformer' object has no attribute 'pos_embedding'"
It seems that timm model is not compatible with the MAE implementation. Can this be easily fixed or I will have to change the internal implementation of MAE?
The text was updated successfully, but these errors were encountered: