USBhost

Results 106 comments of USBhost

This sounds useful. So I could mass convert epubs to text then just dump all the txt of them in a folder and just train off that.

I would like to report all of [Neko's](https://huggingface.co/Neko-Institute-of-Science) tokenizers are current and match with https://huggingface.co/oobabooga/llama-tokenizer. Also if you want me to update stuff in the future just bug me here...

Does it also help the other K quants?

> @USBhost Unfortunately no. The K quants were designed to exploit under-utilization of CPU resources when doing matvecs. I tried copying and pasting the `Q5_K_M` code into a tinyBLAS 2-d...

> It looks like the tests which are currently failing are unrelated to the LLaMA code, so this should be good to review/use. > > If folks can try it...

After replacing transformers from Kobold with this PR I am able to load the shards as expected. Just I cant generate anything because Kobold still needs some changes. ![image](https://user-images.githubusercontent.com/7269941/222938895-2e8b9d71-6a88-417d-b7ed-14d8216d2ef4.png)

KoboldAI now works

So I guess as the first tester. So this is my dataset running at batch 8. ``` Buckets: 384x768: 1 448x768: 1 512x768: 153 768x512: 12 ``` So Currently in...

Everything's playing with LoRAs and they don't see this lol.

This is so amazing. Cooking a batch of 20 at 768x768 at 11250MiB / 12288MiB VRAM usage. I think I could push it to 23. cmdline: --listen --xformers --no-half-vae --deepdanbooru