kusoge

Results 23 comments of kusoge

You can retry the download now, it seems to have fixed itself (at least on my end). I temporarily closed the issue on the new fork release since it's working...

It runs fine on my system. GPU RTX 3060 12GB 16GB RAM CPU Ryzen 5 5600G

Tick the CPU option, it's gonna be painfully slow though.

I managed to run the ggml-alpcca-13B model quite easily. Hopefully it will be supported soon.

When you loaded the webui you are probably on [instruct] mode. ![image](https://user-images.githubusercontent.com/12083007/233371793-f4ca5925-9cb8-4acc-bab1-53083bc840de.png) If you click a character profile in that mode, it throws that error. Just switch to chat

What are your parameter settings? IF you set the max-new-token to 2000(default maximum-unless you changed the max if you were using the MPT model), the AI will ignore context/chat history.

Check ooba's post on reddit. He mentioned something about making it not necessary to go through command-line parameters and improve the UI experience. Perhaps, a fix should be to load...

![image](https://user-images.githubusercontent.com/12083007/231760586-d7c9623a-39c6-4d3a-b933-81a4b221903b.png) You can manually load the model now from this tab

Models that outputs gibberish stuff like that for me are: Neko-Institute_LlaMA-13B-4bit-128g TheYuriLover_llaMA-13b-pretrained-sft-do2-4bit-128g-TRITON Bradaar_toolpaca-13b-native-4bit-128g-cuda Braddar_gpt4-x-alpaca-native-4bit-128g

I'm also getting gibberish response from the newer vicuna-13b-1.1-gptq-4bit-128g model. Apparently, according to the uploader: The safetensors model file was created with the latest GPTQ code, and uses --act-order to...