BLIP icon indicating copy to clipboard operation
BLIP copied to clipboard

Some doubts about weights

Open SKBL5694 opened this issue 3 years ago • 2 comments

I use train_vqa.py to finetune the origin weights "model_base_vqa_capfilt_large.pth" which file size is about 1.34GB. But when I finish the finetune, the new weight is about 4.04GB. Both weights can be loaded by the model. What is the reason for the difference between the two weights?

SKBL5694 avatar Sep 21 '22 01:09 SKBL5694

Hi, for the VQA task, the text encoder and decoder do not share parameters (their parameters are shared during pre-training).

LiJunnan1992 avatar Sep 21 '22 23:09 LiJunnan1992

Hi, for the VQA task, the text encoder and decoder do not share parameters (their parameters are shared during pre-training).

Thanks for your reply. I think you mean the chapter 4.4 and fig.5 in your paper. But I still have some confusion on it. I think I don't fully understand your paper. I want to know if there is a method I can finetune the model by VQA task, but keep the parameters between encoder and decoder shared? Another question is don't you get the original weights by fine-tuning the model on VQA task? If you fine-tune it by VQA, why the parameters between encoder and decoder keep the same; if you don't fine-tune, why the origin weights can perform well during my test on VQA task? If you think the answer is in your paper, please tell me which part of the paper, I will read it carefully again. I'm sorry I didn't fully understand your paper just ask here, looking forward to you giving me some hints again.

SKBL5694 avatar Sep 22 '22 02:09 SKBL5694