bavi!
Results
1
comments of
bavi!
Hi! I can take this. I reproduced the discrepancy with the provided snippet. The gap stems from tokenization differences: evaluate’s BLEU uses SacreBLEU-style tokenization (e.g., 13a), while pycocoevalcap uses COCO’s...