kossum

Results 8 comments of kossum

I added a gemma3 chat handler, it might bring some help. PR: #1989

Sorry, i didn't modify the origin chat template of gemma3 and then used `"type": "image"`. Now i have changed the format of the messages to be compatible with the openai...

Hello @joaojhgs, gemma3 (especially the 12b and 27b versions) has strong instruction-following abilities and can generate structured function call outputs through well-designed prompts. But unlike gpt4 or claude, gemma3 does...

Hello @Domino9752 ,thanks for testing! The issue happens because the author updated the llama.cpp library in the current 0.3.9 version, but the corresponding changes for the llava part haven’t been...

Hi @xia0nan, thanks for your feedback. If you find that inference speed is slow, please note that you can use the `n_gpu_layers` parameter to specify how many transformer layers should...

Hi @Gordonei , thanks for your message and for the detailed feedback! I haven’t logged into github for a while, so I wasn’t aware of the recent updates. I’ll rebase...

Updated to 0.3.14. Since llama.cpp’s mtmd module now covers the image embedding for gemma3, I’ve removed my previous implementation and am now only keeping the chat_template.

This library does not currently support the multimodal features of gemma3. For this functionality, please see PR #1989