shidingz
shidingz
minigemini_instruction.json中有条目的image为llava/LLaVA-Pretrain/images/00013/000133305.jpg,有很多llava的图片,但是readme中没有提到要用llava的图片,是你们疏忽了吗
Traceback (most recent call last): File "/checkpoint/binary/train_package/minigemini/train/train_mem.py", line 14, in train(attn_implementation="flash_attention_2") File "/checkpoint/binary/train_package/minigemini/train/train.py", line 1262, in train trainer.train() File "/root/.local/lib/python3.8/site-packages/transformers/trainer.py", line 1624, in train return inner_training_loop( File "/root/.local/lib/python3.8/site-packages/transformers/trainer.py", line 1961,...
https://github.com/dvlab-research/MGM/blob/main/scripts/llama/train/stage_1_2_full_v7b_336_hr_768.sh 在这个脚本中pretrain用的--version plain 而finetune用的是--version v1 前后不一致模型不会混乱吗
I found something strange when loading the model. It seems that you have released the vision_tower during training, but when loading the vision_tower, you did not load the gradient-updating parameters,...
When I run this script-scripts/llama3/train/stage_2_full_v8b_672_hr_1536.sh, I encounter this error- WARNING: tokenization mismatch: 156 vs. 161. (ignored)