LWM icon indicating copy to clipboard operation
LWM copied to clipboard

`model_max_length` why is it 2048?

Open denadai2 opened this issue 1 year ago • 1 comments

Hi everyoneee,

I anticipate that it might be a stupid question but why do we have model_max_length: 2048 in the tokenizer_config.json https://huggingface.co/LargeWorldModel/LWM-Text-Chat-1M/blob/main/tokenizer_config.json?

Thank you!

denadai2 avatar Feb 28 '24 16:02 denadai2

I encountered similar problem. When I use LWM-TEXT-512K (pytorch), warning that "Token indices sequence length is longer than the specified maximum sequence length for this model (42314 > 2048). Running this sequence through the model will result in indexing errors". When the input length exceeds 40k, the model will produce confusing output results:"['\n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n### \n\n##']"

LzhinFdu avatar Mar 06 '24 08:03 LzhinFdu