AGI-player

Results 14 comments of AGI-player

> ## Why are these changes needed? > [TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM) can greatly improve the inference speed of LLM. It would be helpful to support tensorRT-LLM in Fastchat. > > This commit...

> 请问你是如何使用llama 13b的config,然后随机初始化权重的,启动要设置什么参数? 我也想尝试pretrain 参考的https://github.com/FlagAlpha/Llama2-Chinese/blob/main/train/pretrain/pretrain_clm.py 需要在LLaMA-Factory对应加载模型的地方进行修改 model = AutoModelForCausalLM.from_config(config) n_params = sum({p.data_ptr(): p.numel() for p in model.parameters()}.values()) logger.info(f"Training new model from scratch - Total size={n_params/2**20:.2f}M params")

> > > 请问你是如何使用llama 13b的config,然后随机初始化权重的,启动要设置什么参数? 我也想尝试pretrain > > > > > > 参考的https://github.com/FlagAlpha/Llama2-Chinese/blob/main/train/pretrain/pretrain_clm.py 需要在LLaMA-Factory对应加载模型的地方进行修改 > > ``` > > model = AutoModelForCausalLM.from_config(config) > > n_params = sum({p.data_ptr(): p.numel() for p...

> @AGI-player Thanks for your feedback. This is a known issue and we will fix it soon. ok~

> > @AGI-player Thanks for your feedback. This is a known issue and we will fix it soon. > > ![image](https://private-user-images.githubusercontent.com/16505966/331958695-bb56c009-3c74-40fd-912a-ba5ef8a95b77.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTYxODUyOTUsIm5iZiI6MTcxNjE4NDk5NSwicGF0aCI6Ii8xNjUwNTk2Ni8zMzE5NTg2OTUtYmI1NmMwMDktM2M3NC00MGZkLTkxMmEtYmE1ZWY4YTk1Yjc3LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA1MjAlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNTIwVDA2MDMxNVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTlmZTg1OTM2MGQ2YTNjOTJlNzM4OTM0OTkyNDUwZTMzY2JlZjVjY2IzMDA4MDUzM2UyOTFjNmU0NmUxNDE3YjkmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.dPTMOkx7ydWPI_ojfrUuTgLN7sNZ9tLOY1T4MM-a7kI) I got this error it should be caused by...

> > > > @AGI-player Thanks for your feedback. This is a known issue and we will fix it soon. > > > > > > > > > ![image](https://private-user-images.githubusercontent.com/16505966/331958695-bb56c009-3c74-40fd-912a-ba5ef8a95b77.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTYxODUyOTUsIm5iZiI6MTcxNjE4NDk5NSwicGF0aCI6Ii8xNjUwNTk2Ni8zMzE5NTg2OTUtYmI1NmMwMDktM2M3NC00MGZkLTkxMmEtYmE1ZWY4YTk1Yjc3LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA1MjAlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNTIwVDA2MDMxNVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTlmZTg1OTM2MGQ2YTNjOTJlNzM4OTM0OTkyNDUwZTMzY2JlZjVjY2IzMDA4MDUzM2UyOTFjNmU0NmUxNDE3YjkmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.dPTMOkx7ydWPI_ojfrUuTgLN7sNZ9tLOY1T4MM-a7kI)...

> @AGI-player Thanks for your feedback. This is a known issue and we will fix it soon. @jershi425 is this issue fixed in the latest version?

> Please follow the issue template to share the full end to end reproduced steps. Thank you for cooperation. the trt engine was built with: trtllm-build --gemm_plugin float16 --max_batch_size=128 --max_input_len=8192...

> The `stop_words_list` is not supported well in 0.9.0, maybe you can try the latest main branch, we have refactored the `GenerationExecutor`, and the [stop_words](https://gitlab-master.nvidia.com/ftp/tekit/-/blob/main/tensorrt_llm/executor.py#L106) are supported. I update the...

> Hi @AGI-player , what's tensorrt_llm version? Could you please try the main branch? apt-get update && apt-get -y install python 3.10 python3-pip openmpi-bin libopenmpi-dev git git-lfs pip3 install tensorrt_llm...