Guaguago
Guaguago
Are those samples in the [ human_anotation/pplm_labled_csvs ] directory?
@ehsan-soe hi ```python import math import torch from transformers import OpenAIGPTLMHeadModel, OpenAIGPTTokenizer tokenizer = OpenAIGPTTokenizer.from_pretrained('openai-gpt') model = OpenAIGPTLMHeadModel.from_pretrained('openai-gpt') model.eval() def score(sent): indexed_tokens = tokenizer.encode(sent) tokens_tensor = torch.tensor([indexed_tokens]) with torch.no_grad(): outputs...
@ehsan-soe do you know how can I use this code to get the perplexity scores of paper?
@dathath Sorry I can't find any samples generated in this repository, can you give me the specified location or some instructions on how can I use this code to get...
@ehsan-soe Soga! Thank you! @dathath Thank you! and I find that each item in the CSV file has 2 generated samples. It seems that one sample is from PPLM and...
@dathath Thank you very much and I will try it! And I have made two programs to test PPL and distinct-n respectively according to your suggestions before. But the scores...
@dathath Thank you so much! Really helpful clue by which I have got the exactly same Dist-1,2,3 scores as paper. But for PPL, most of the scores I have got...
@dathath Is there any special process to the token "" and the '\n' within a sentence in the calculation of PPL? Should I drop them before calculating ppl?
@dathath After having solved some bugs and warnings, I found that my results of PPL turn to **much lower** than the paper but the dist scores almost **ideally match**. I...
@dathath @Andrea Hi, thank you, you are so nice! This is my email: [email protected] I need your help, please!