Lu Mingcong
Lu Mingcong
it's because poisson disstribution has an infinite upper limit, so the author just take a limit of 10 cars. what you said about execute 0-20, it is reflected in this...
What's your torch version. I use torch 2.0 at first and got same problem, then I degraded it to 1.13.1 and works well. Hope helpful
> Hello, was just trying this out as well; Using the latest `peft` as suggested gets rid of the "cannot flatten integer dtype tensors" error. However, a new error now...
> Tried these > > ``` > pip uninstall peft > pip install git+https://github.com/huggingface/peft.git > ``` > > and getting the same error: > > ``` > rank1]: File "...../LLaMA-Factory/v/lib/python3.11/site-packages/peft/tuners/lora/dora.py",...
Same problem in llamafactory(latest pull 2025.4.29) https://github.com/open-thoughts/open-thoughts/issues/30
There is indeed no zero_pp_rank_4_mp_rank_00_optim_states.pt in .../checkpoint-1500/global_step1499/. What can I do ``` $ls checkpoint-1500/global_step1499/ bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt bf16_zero_pp_rank_10_mp_rank_00_optim_states.pt bf16_zero_pp_rank_20_mp_rank_00_optim_states.pt bf16_zero_pp_rank_30_mp_rank_00_optim_states.pt bf16_zero_pp_rank_11_mp_rank_00_optim_states.pt bf16_zero_pp_rank_21_mp_rank_00_optim_states.pt bf16_zero_pp_rank_31_mp_rank_00_optim_states.pt bf16_zero_pp_rank_12_mp_rank_00_optim_states.pt bf16_zero_pp_rank_22_mp_rank_00_optim_states.pt bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt bf16_zero_pp_rank_13_mp_rank_00_optim_states.pt bf16_zero_pp_rank_23_mp_rank_00_optim_states.pt bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt bf16_zero_pp_rank_14_mp_rank_00_optim_states.pt...
> try --enable_liger_kernel and --use_unsloth_gc --use_unsloth_gc or --use_unsloth?
> use_unsloth_gc Thanks. BTW, I have encounter an error : Triton Error [CUDA]: device kernel image is invalid when --enable_liger_kernel. Here are some pkg info: triton==3.1.0 transformers==4.44.2 torch=2.3.0 CUDA SDK...
请问你72b,16k长文本用的多少资源呢?8卡80g能行吗