Xinrong Zhang
Xinrong Zhang
We have re-uploaded the files and solved this problem. Please kindly use load_dataset("xinrongzhang2022/InfiniteBench") now!
please kindly set ``features`` when loading datasets ``` import datasets from dataset import Value, Sequence ft = Features({"id": Value("int64"), "context": Value("string"), "input": Value("string"), "answer": Sequence(Value("string")), "options": Sequence(Value("string"))}) dataset = load_dataset("xinrongzhang2022/InfiniteBench",...
from datasets import load_dataset, Features, Value, Sequence ft = Features({ "id": Value("int64"), "context": Value("string"), "input": Value("string"), "answer": Sequence(Value("string")), "options": Sequence(Value("string")) }) dataset = load_dataset("xinrongzhang2022/InfiniteBench", features=ft)
We have revised the En.QA task. And those two models are evaluated at different task versions
We have fixed this bug in the latest commit. Thank you for your notification.
会存在这样的问题
it seems that is a special token, and you can pre-process it
huggingface data数据集是有更新的,以数据集为准
the context is long and noisy. LLMs tend to "think" rather than guessing. The output is also posted under our repo
if a model generates none, it will get zero at this example