LWShowTime

Results 17 comments of LWShowTime

> I found that when commented out the line in /model/blip.py line 131 fix the problem: > > ![image](https://private-user-images.githubusercontent.com/57023743/287519088-bf843a0a-a070-4dde-ab87-db0cb25da79c.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MDQ2NDExMjMsIm5iZiI6MTcwNDY0MDgyMywicGF0aCI6Ii81NzAyMzc0My8yODc1MTkwODgtYmY4NDNhMGEtYTA3MC00ZGRlLWFiODctZGIwY2IyNWRhNzljLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDAxMDclMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwMTA3VDE1MjAyM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWViNzI5YmEyMzAyODlkNzkyNTNhMGM3NmNhYjg3M2UzYjVhOWEzZTMyMGM1NGY4ZDAyNTc0MDRjNWM1MTM3YmEmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.7-OQTQIEJPKQPxqBquhAneH9SOPuBwDmrgW_kPQLofk) > > Don't know why, hope someone can provide the detail...

I solved this problem, if the transformers is 4.16.0, everything is ok. But I used the transformers4.36.2, in this case, in the 818 lines of the generation_utils.py of transformers: you...

You guys can run the sample code? Where did you find the dataset?

My launch : --version=Mylocalpath/LISA/ --vision_tower=Mylocalpath/CLIP-vit-large-patch14/ --precision=fp16 --load_in_4bit

![image](https://github.com/dvlab-research/LISA/assets/51390653/c74d0425-cb3c-4248-a5f9-c4a1f1752a04)

Actually, this is the bug of the deepspeed, I think you can avoid this by not using fp16. Try use bf16 ir fp32 @ZichengDuan

> @ZichengDuan Same situation as you, so I chose a GPU with 32G VRAM, and everything goes on well LOL. But the problem of dim1 > dim2 when using 4bit...

@shell-nlp Have been solved if you use a GPU with 32G VRAM.

![image](https://github.com/dvlab-research/LISA/assets/51390653/94b5ab8d-a941-4a31-937a-ae7947d4a780) Is this warning about bfloat16 matters? In the last version I remember the inference is fast when segment an image which took only severl seconds. However, the new 13B...

@X-Lai And when I run version 2 in the multi-gpu devices. I got this error: `indices should be either on cpu or on the same device as the indexed tensor...