Liya Wang

Results 15 issues of Liya Wang

Hello, I downloaded youtube vis 2019 dataset. The format is different for your required format. Therefore, i am wondering why we used your linked website and got different formats. Is...

I used your fastconvmae train imgnet data. In your code, you said the output should be: ![image](https://user-images.githubusercontent.com/41125918/183908655-4b73bfa0-7c7a-442e-b7e9-2b4db42da3be.png) However, when i used the pretrained model to predict, it gave me prediction...

I downloaded your pretrained model. And when i tried to load it. It gave me the following errors. _IncompatibleKeys(missing_keys=['mask_token', 'decoder_pos_embed', 'stage1_output_decode.weight', 'stage1_output_decode.bias', 'stage2_output_decode.weight', 'stage2_output_decode.bias', 'decoder_embed.weight', 'decoder_embed.bias', 'decoder_blocks.0.norm1.weight', 'decoder_blocks.0.norm1.bias', 'decoder_blocks.0.attn.qkv.weight', 'decoder_blocks.0.attn.qkv.bias',...

Do you have codes for linear probe?

In your method @register_model def cae_large_patch16_384(pretrained=False, **kwargs): model = VisionTransformer( img_size=384, patch_size=16, embed_dim=1024, depth=24, num_heads=16, mlp_ratio=4, qkv_bias=True, norm_layer=partial(nn.LayerNorm, eps=1e-6), **kwargs) model.default_cfg = _cfg() return model def _cfg(url='', **kwargs): return {...

I test SeqFormer model with res = Trainer.test(cfg, model). And the output is empty, and I debug it. And found the followings from seqformer.py line 238. {'image_size': (720, 1280), 'pred_scores':...

Hello, can you provide the pre-trained models of SeqFormer? Thanks,