Seungyoun, Shin

Results 18 comments of Seungyoun, Shin

you are right. I have big mistake. I will update it as soon as possible.

I dont know why this error occured. It's probably different torch version creating zero_tensor in different way. I will check it as soon as possible

This is the result from "COCO 2014 train" , almost the same dalle structure @ 1epoch I notice that after 1epoch loss doesn't seem to decrease and the image is...

Do we have train code for hf-owl-vit?

batch size 1 and gradient accumulation 1 works for more than 2000 steps. Not sure I can finish training without OOM

I also posted https://github.com/tatsu-lab/stanford_alpaca/files/11024692/Mar20_05-17-08_0c56f6779a08.csv csv log file! I takes approx. 24hours(a day) This is strange. You are using way better gpus than mine. As you mentioned wandb could be the...

--bf16 True \ --per_device_train_batch_size 1 \ --per_device_eval_batch_size 1 \ --gradient_accumulation_steps 1 \ @danwei1992 Make sure this 4 terms are correct I am using - docker image :[hf:latest]( https://hub.docker.com/r/huggingface/transformers-pytorch-gpu) - CUDA...

This is my experiments with prompt tunning. Not perfect but pretty amazing Seems like img1,img2,text is better performing.

![output](https://user-images.githubusercontent.com/20262536/225847888-e3611696-bc00-4289-a4a8-fb97b3d89ae6.png) Deleting the line and build from source works. I am not sure this is desired or not + At ProcThor, we cannot goto some locations even if the location...

\I have successfully modified the Unity code in Fill.cs to disable the spilling action. However, I am still uncertain about why the liquid spills when teleporting without tilting the held...