Niranjan Akella
Niranjan Akella
@EricLBuehler Non-Quantized f16,32 implementation currently holds more precedence. But if possible, would also like to have a quantized implementation too. Also I wish to know if LoRA adapters can be...
@EricLBuehler No problem sounds good. I am looking forward to trying it out soon.
Hey the tokenizer.model is still missing from the official repo, unable to convert tl2 without it. @BradZhone
@minjikarin @cbetta @anthonyroux Please let me know how I can solve this or if I am making some mistake in the way I am using this API.
@freddyaboulton Any idea how to host a local or bedrock LLM for this purpose like I am trying to do? Should have have a different session created with the audio...