WangXu

Results 5 issues of WangXu

>alibi捕获相对位置,对角线元素应该一致吧,但是为什么baichuan的不一致?是有什么玄机吗? ![Snipaste_2023-09-14_17-10-25](https://github.com/baichuan-inc/Baichuan-13B/assets/31993091/9bba8090-51ab-454e-bc9d-279c74f301b2)

I used the example script in the readme to quantize llama3-8b ``` python quant_config = { "zero_point": True, "q_group_size": 16, "w_bit": 4, "version": "GEMM" } model = AutoAWQForCausalLM.from_pretrained(model_path, **{"low_cpu_mem_usage": True},...

`AttributeError: 'LlamaTransformerLayerWeight' object has no attribute 'q_weight_'` The 4bit version seems to be only "att_norm_weight_" and "ffn_norm_weight". When will the 4bit version be supported?

您好, 我使用单块A800进行部署推理时正常,但是使用多卡推理会报错: `Task exception was never retrieved future: Traceback (most recent call last): File "/opt/conda/lib/python3.9/site-packages/rpyc/core/stream.py", line 268, in read buf = self.sock.recv(min(self.MAX_IO_CHUNK, count)) ConnectionResetError: [Errno 104] Connection reset by peer...

我严格按照README安装了相关的包 pip install -r requirements.txt pip install git+https://github.com/unslothai/unsloth.git pip install bitsandbytes==0.43.1 pip install peft==0.10.0 pip install torch==2.2.2 pip install xformers==0.0.25.post1 启动参数 ```shell export CUDA_VISIBLE_DEVICES=2,3,4,5,6,7 torchrun --nproc_per_node=6 train.py \ --train_args_file train_args/sft/qlora/llama3-8b-sft-qlora.json...