XPretrain
XPretrain copied to clipboard
Multi-modality pre-training
Is the line in 'lfvila8m_clipid.jsonl' a video clips-sentence pair? And I see an variational number of video-clips per row. So how the video-clips of 'lfvila8m_clipid.jsonl' is divided from the original...
May I request the transcript text processing code? [[email protected]](mailto:[email protected]) Thank you very much!
Thanks for your great work! I want to follow your work, but I meet some problems with the dockfile. It seems the image nvidia/cuda:10.1-devel-ubuntu18.04 not exist. Can you provide a...
the padding seems not right, or maybe i made a mistake ``` # padding _, _, D, H, W = x.size() if H % self.patch_size[0] != 0: x = F.pad(x,...
Update Dockerfile for Issue #34
When we run `horovodrun -np 1 python src/pretrain/run_pretrain.py --config src/configs/msrvtt_retrieval/msrvtt_retrieval_vip_base_32.json` We get the following error: ```:Traceback (most recent call last): : File "src/pretrain/run_pretrain.py", line 22, in : from transformers import...
First of all - Amazing work on this one. I'm a bit getting lost with the repo, may I request a simple few line script that does something like the...
Bumps [transformers](https://github.com/huggingface/transformers) from 4.30.0 to 4.36.0. Release notes Sourced from transformers's releases. v4.36: Mixtral, Llava/BakLlava, SeamlessM4T v2, AMD ROCm, F.sdpa wide-spread support New model additions Mixtral Mixtral is the new...
Excellent work! I'm trying to deploy some attacks on your models, but I cannot fine-tune your pretrained ones on my local server due to a VRAM shortage. Could you please...
When running the command inside the docker image for finetuning LF-VILA, following error is created, root@8dccc81930c3:/LF-VILA# deepspeed src/tasks/run_video_classification.py --distributed --blob_mount_dir /blob_mount --config $CONFIG_PATH --deepspeed [2023-10-17 11:11:02,765] [WARNING] [runner.py:132:fetch_hostfile] Unable to...