ScaleLLM icon indicating copy to clipboard operation
ScaleLLM copied to clipboard

Quantization: Supporting FP8 for both models and KV caches

Open guocuimi opened this issue 1 year ago • 0 comments

guocuimi avatar Apr 28 '24 04:04 guocuimi