vidhyat98
vidhyat98
when can we expect awq models to be optimized for inference?
any updates on this?
@SalmanMohammadi When can we expect the cookbook to be updated? Would the files still work if vllm functionality is not used? My training doesn't seem to converge.
Facing the same issue with llama 3.1 model adapters
I'm facing the same issue. ETA on when the fix would be merged?
Observing the same issue while trying to serve on 2 A100 40GB GPUs.