Results 10 comments of cOng

Hi @christobill, could you help to test with your models?

> I think llama2 also have the same situation. So for llama2` ` is the stop word? Test on LLAMA2, The prompt: `[INST] hi. Evaluate translation from English #This section...

HI, I mean the official stop tokens is indeed ``, but as you can see the conversation class has defined llama2 template with ` ` as stop string, for compatibility,...

Hi @infwinston , this PR is ready to be merged, could you help to have a final reviews and merged, so as the associated doc PR https://github.com/lm-sys/FastChat/pull/3139 These features should...

> @congchan Could you add the deepspeed zero3 support on the triain_with_template? > > Do you think if it should add > > ``` > if trainer.is_deepspeed_enabled: > trainer.save_model() >...

It is actually 1024. Can the team share guides to fine-tune 8K context length?

Hi, may I suggest to support **copy** the whole conversation history as well?

Run into the same issue for normal decoder only model

Hi guys, I encountered the same error for version 4.41.2. I am confused which package dose `flash_attn` refered to? I have tried install both of the two packages (xformers and...

Hi, I encounter similar issues, I am not sure if this is related with HF, since the value error arise for vllm>=0.7.3