wanda icon indicating copy to clipboard operation
wanda copied to clipboard

llama_7b wikitext perplexity 7.0915350914

Open xiaopengaia opened this issue 1 year ago • 3 comments

bash scripts/llama_7b.sh

the source model: wikitext perplexity is 5.67702

prune this model, sparsity 50%, get wikitext perplexity is 7.09153509

but the paper is : 50% 7.26

why?

xiaopengaia avatar Apr 30 '24 08:04 xiaopengaia

Are you using the c4 calibration dataset or the redpajama? In my experience, using the calibration dataset from redpajama could get numbers as high as around 7.10.

Eric-mingjie avatar Apr 30 '24 12:04 Eric-mingjie

Hi. Were you able to resolve the issue?

taratt avatar May 08 '24 17:05 taratt

Thanks, I didn't use c4 calibration dataset, I just wanted to reproduce the data in the paper.

def get_loaders(name, nsamples=128, seed=0, seqlen=2048, tokenizer=None):
    if 'wikitext2' in name:
        return get_wikitext2(nsamples, seed, seqlen, tokenizer)
    # if "c4" in name:
        # return get_c4(nsamples, seed, seqlen, tokenizer)

xiaopengaia avatar May 09 '24 03:05 xiaopengaia