ShahZ181

Results 1 issues of ShahZ181

Has anyone gotten 16k context length with codellama or llama2? because i have tried multiple models but they all start producing gibberish when the context window gets past 4096. I...