Chancharik Mitra
Chancharik Mitra
### Question Hello! I am curious to run an experiment using LLaVA-1.5 without images (to clarify; it is important for my work to specifically prompt LLaVA-1.5 without images _not_ Vicuna13B)....
Hello, Brady! Thanks for compiling so many great methods into this very helpful resource. Our paper is a multimodal CoT method that has been out for a little while and...
I have successfully visualized individual tissue samples with a standard output structure using `SpatialFeaturePlot`, but I would like to visualize multiple samples in one of Spaceranger's _aggregated outputs._ The output...
Hello! First of all, this is really fascinating work. Thanks for the contribution. I wanted to reach out and ask if you could share the evaluation scripts for mPLUG-OWL2 you...
First of all, thanks for the contribution! I was interested in probing the results of your method on datasets like SEEDBench and MMBench. Do you have any scripts, special prompting...
Hello, and thanks for such a great contribution to the field of interleaved LMMs! This is really great work. I was wondering if there was an example of the format...
Hi, thanks for your fantastic video foundation model! I was interested in exploring the capabilities of InternVideo2-Chat for both images and video. According to the Huggingface code, the model can...
Hi, simple question about your fantastic model! I see that your multi-image demo uses image tokens while your single high-resolution image demo does not. Is the usage of image tokens...
This PR includes the following additions: 1. Support for video VQAScore inference for all previously-supported T2V models. 2. Addition of new interleaved-image/video LMM models. 3. Updating of the README and...
Hello, thanks for contributing a very exciting model! I noticed that the interleaved and video inference examples given in the notebooks are set up with different configs as the model...