yiyexy
yiyexy
hello everybody, I met the same problem. And finally I found the key of this problem. If you set your workers_per_gpu as 0, you will get the error log same...
I found that the reason for this problem is different tokenizer rules. The `bos_token` is null and the `eos_token` is set to "" in the Qwen tokenizer configuration. So I...
> I found that the reason for this problem is different tokenizer rules. The `bos_token` is null and the `eos_token` is set to "" in the Qwen tokenizer configuration. So...
> @yiyexy hello, nice catch. Am training normal now. Did u trained on llava pretrain data? Does there any pretrain data could be used for Chinese enhancement ? Yes, I...
@lucasjinreal I will. But it still has some problems to be solved. It's a long way.
> @yiyexy Hello, Your loss looks not like stage 1? > > BTW, you probably should use qwen1.5-7b-chat model. Otherwise you can not sft efficiently. > > However, qwen using...
@20191864218 Maybe you need set some parameters for Qwen1.5. #1146
> @yiyexy Using llava template on qwen chat model might introduce unwanted output when chat. This is a common issue. qwen using chatml format which using as spepartor/ Thanks for...
> > I think base can not be used in vlm, it doens't have chat abilities,. > > I want to create a model solely for generating reports, without requiring...