VBench icon indicating copy to clipboard operation
VBench copied to clipboard

human_action question

Open zixian-wu opened this issue 1 year ago • 4 comments

I've tried many times and the human_action score stays at 0. Can you give me some ideas? { "human_action": [ 0.0, [ { "video_path": "./sampled_videos/anime-new-vbench/0001.mp4", "video_results": false, "cor_num_per_video": 0 }, { "video_path": "./sampled_videos/anime-new-vbench/0002.mp4", "video_results": false, "cor_num_per_video": 0 }, { "video_path": "./sampled_videos/anime-new-vbench/0003.mp4", "video_results": false, "cor_num_per_video": 0 }, { "video_path": "./sampled_videos/anime-new-vbench/0004.mp4", "video_results": false, "cor_num_per_video": 0 }, { "video_path": "./sampled_videos/anime-new-vbench/0005.mp4", "video_results": false, "cor_num_per_video": 0 } ]

zixian-wu avatar Oct 25 '24 02:10 zixian-wu

Mabey because the video_path is not correct. I was trying to evaluate some videos generated by myself and I met the same problem. Then I read the code in vbench/human_action.py, I found they may store action information in the file name. And they detect the actions in the video, compare them with the target action which appears in the file name.

LazySheeeeeep avatar Nov 08 '24 10:11 LazySheeeeeep

i also find this question, it shows that the encode embedding of the video is 0, so strange

DidiD1 avatar Dec 03 '24 07:12 DidiD1

Thanks for your interest in our work! For the human action dimension, since the classifier categories are fixed, we do not support customized video evaluation yet, and the name of the sampled video should be the same as the prompt for evaluation when using our provided prompt list.

JustinYuu avatar Dec 30 '24 02:12 JustinYuu

so how to compute the long-prompt's human action?

huangjch526 avatar Jan 24 '25 02:01 huangjch526

VBench's human_action dimension currently operates only on the predefined benchmark prompts.

If you're referring to augmented prompts, you can still use the same evaluation pipeline—just apply it to the longer prompts derived from the human_action short prompts.

If you're referring to open-domain prompts, that’s not supported yet in this dimension. However, feel free to explore VBench-2.0, which offers expanded evaluation options for motion rationality and human fidelity.

Thanks!

ziqihuangg avatar May 06 '25 08:05 ziqihuangg