怪猎糕手
怪猎糕手
root@c68c31f45482:/workspace/zt/code/FastChat# python3 -m fastchat.model.apply_delta --base-model-path ../../model/Llama-2-7b-hf --target-model-path ../Sequence-Scheduling/ckpts/vicuna-7b --delta-path lmsys/vicuna-7b-delta-v1.1 Loading the delta weights from lmsys/vicuna-7b-delta-v1.1 You are using the default legacy behaviour of the . This is expected, and...
Hi! I am a graduate student who have read the paper about the Deepspeed framework.[https://arxiv.org/pdf/2207.00032]() I gain great interests on the hybrid pipline inference strategy descrided in Section Ⅳ, but...
Hi, when I try to run the `pippy_llama.py` in this repo, it show that there is a bug: ``` root@6e61f182b97b:/zt/code/my_dev# torchrun --nproc-per-node 4 pippy_llama.py W1027 12:28:26.326000 25180 torch/distributed/run.py:793] W1027 12:28:26.326000...
Hi, I am a student who interested in pipeline parallelism for LLM inference. I have successfully run the [example](https://github.com/pytorch/PiPPy/blob/main/examples/cpu_init/gpt2_cpu_init.py) for GPT mentioned in Pytorch document, so I just want to...