ConvMAE
ConvMAE copied to clipboard
ConvMAE: Masked Convolution Meets Masked Autoencoders
I used your fastconvmae train imgnet data. In your code, you said the output should be:  However, when i used the pretrained model to predict, it gave me prediction...
I downloaded your pretrained model. And when i tried to load it. It gave me the following errors. _IncompatibleKeys(missing_keys=['mask_token', 'decoder_pos_embed', 'stage1_output_decode.weight', 'stage1_output_decode.bias', 'stage2_output_decode.weight', 'stage2_output_decode.bias', 'decoder_embed.weight', 'decoder_embed.bias', 'decoder_blocks.0.norm1.weight', 'decoder_blocks.0.norm1.bias', 'decoder_blocks.0.attn.qkv.weight', 'decoder_blocks.0.attn.qkv.bias',...
What should I do if I want to fine tune the current pre training model to my own dataset instead of Imagenet's Val dataset? Can you answer it? Thank you...
Thank you for your excellent work. I noticed that in Table 6, the results of LIN pretrained with ConvMAE-Based for 200 epochs are missed. May I ask what they are?...
Hi, I am looking to see how well the pretrained base model runs on my own dataset, but the current model is configured for an image size of 224 In...
Thank you for your impressive work! VideoConvMAE seems still lack a code release, can you update it?