UltraChat icon indicating copy to clipboard operation
UltraChat copied to clipboard

Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)

Results 22 UltraChat issues
Sort by recently updated
recently updated
newest added

Hi team, Thank you for the great work. I'm wondering if you can post the prompt to interact with user's question during data collection. I saw the system prompt for...

I observe that the inference script contains an embedded system prompt. Is it also included in training? I wonder how much the performance will be affected if changing the system...

greate job, Tks you so much for sharing, I have a question about prompt engineering that You make a dialog around 1 topic like you say: "We sample 200k specific...

Hello! Thanks for the awesome work! I am trying to find the metadata, e.g. topic, corresponding to each data point. They don't seem to be present in the current released...

> We instruct the user model with carefully designed prompts to mimic human user behavior @ningding97 could you provide these prompts you used? Thanks!

你好,非常感激开源训练数据。 我想请教下你们训练13b模型使用的超参是多少?训练了多少步数?

从hugging face上下载了模型 运行chat_cli.sh 模型回复出现乱码问题 这是为什么? ![image](https://github.com/thunlp/UltraChat/assets/72975999/0a6851d7-491e-46b5-b9ec-95044ebc326f)

def get_model_tokenizer(args): model = LlamaForCausalLM.from_pretrained(args.model_name_or_path) tokenizer = LlamaTokenizer.from_pretrained(args.model_name_or_path) tokenizer.add_special_tokens({'pad_token': ""}) model.resize_token_embeddings(len(tokenizer)) model = bmt.BMTrainModelWrapper(model) return model, tokenizer 假设在单机8卡服务器上,加载UltraChat65B的模型进行微调,会不会存在OOM的问题?每个卡都会执行model = LlamaForCausalLM.from_pretrained(args.model_name_or_path)加载一份模型,哪怕存CPU内存,65B大概需要130G的内存,8卡差不多需要1T的内存,而服务器总内存也差不多1T。

sh-4.2$ sudo sh recover.sh File "weight_diff.py", line 23 special_tokens_dict: Dict, ^ SyntaxError: invalid syntax