Chancharik Mitra

Results 12 issues of Chancharik Mitra

### Question Hello! I am curious to run an experiment using LLaVA-1.5 without images (to clarify; it is important for my work to specifically prompt LLaVA-1.5 without images _not_ Vicuna13B)....

Hello, Brady! Thanks for compiling so many great methods into this very helpful resource. Our paper is a multimodal CoT method that has been out for a little while and...

I have successfully visualized individual tissue samples with a standard output structure using `SpatialFeaturePlot`, but I would like to visualize multiple samples in one of Spaceranger's _aggregated outputs._ The output...

Hello! First of all, this is really fascinating work. Thanks for the contribution. I wanted to reach out and ask if you could share the evaluation scripts for mPLUG-OWL2 you...

First of all, thanks for the contribution! I was interested in probing the results of your method on datasets like SEEDBench and MMBench. Do you have any scripts, special prompting...

Hello, and thanks for such a great contribution to the field of interleaved LMMs! This is really great work. I was wondering if there was an example of the format...

Hi, thanks for your fantastic video foundation model! I was interested in exploring the capabilities of InternVideo2-Chat for both images and video. According to the Huggingface code, the model can...

Hi, simple question about your fantastic model! I see that your multi-image demo uses image tokens while your single high-resolution image demo does not. Is the usage of image tokens...

This PR includes the following additions: 1. Support for video VQAScore inference for all previously-supported T2V models. 2. Addition of new interleaved-image/video LMM models. 3. Updating of the README and...

Hello, thanks for contributing a very exciting model! I noticed that the interleaved and video inference examples given in the notebooks are set up with different configs as the model...