Kevin

Results 30 comments of Kevin

Thanks, We get similar results ~8 [email protected] based on the default settings, and we further boost the performance based on some parameter tuning (e.g., learning rate, batch size). I attached...

Yes, during the feature extraction, the model contains `video_proj` and `text_proj`, and the channels of video and text features are 256.

@takfate Hi, the NLQ results are implemented by my collaborator Mattia, I may misalign some details, I attach our VSLNet code implementation here so that you can refer to the...

Hello Junwei, Thanks for your interest for our work, I will update the instruction and related details for MQ next, Thank you for your patience!

Hi Junwei, I have uploaded the video features for MQ tasks to G drive: [train&val](https://drive.google.com/file/d/1-HEUCdyfNX7CBZhz40yiyTr7to_p7wUi/view) / [test](https://drive.google.com/file/d/1-JmezY3eIkHKJ1JBA_AA8QWBoY3W2HpS/view), so that you can download it directly. What you need to do is...

There is a gz file, after unzipping it (I unzip it on my mac), you will see a document that contains multiple `*.pt`. e.g., `0a8f6747-7f79-4176-85ca-f5ec01a15435.pt`, this pt file corresponding to...

Yes, it is the clip ID. And sorry, I am currently unable to provide video-level features, a solution is to rewrite the data loader so that supports clip features as...

Hi, @srama2512 , I released the codebase here [MQ.zip](https://github.com/showlab/EgoVLP/files/9596196/MQ.zip), you can check the data loader detail regarding clip-level feature loading. Besides, I am able to check the config parameters, can...

Hi, @iranroman I think you do not need to worry about this message, you do not need to fine-tune after loading this checkpoint, directly inference should be able to reproduce...

Hi @thechargedneutron , Currently, this checkpoint should be able to reproduce performance directly with inference, not additional fine-tuning. we have a plan to re-organize our codebase including config, and I...