bavi!

Results 1 comments of bavi!

Hi! I can take this. I reproduced the discrepancy with the provided snippet. The gap stems from tokenization differences: evaluate’s BLEU uses SacreBLEU-style tokenization (e.g., 13a), while pycocoevalcap uses COCO’s...