Melisa Russak

Results 6 comments of Melisa Russak

Hi, I am facing the same problem. Here are my (unsuccessful) attempts: - Set batch size to `1` and tried decoding on CPU. For sentences ~10 tokens I am getting...

> https://github.com/tloen/alpaca-lora > I am keep getting this error message, I am wondering whether you have seen it: `Exception: Could not find the transformer layer class to wrap in the...

> I agree with the second two, but is it bad that > > > `PATTERN_REF` copies only the first matched token > > ? That was intentional. I guess...

Another use case: - matching a token and moving it to the end of the sentence. If you can't copy more than one token - this won't work. It's more...

Hi! I got a similar problem while running the sample code: ```python import torch from transformers import AutoTokenizer from based.models.gpt import GPTLMHeadModel tokenizer = AutoTokenizer.from_pretrained("gpt2") model = GPTLMHeadModel.from_pretrained_hf("hazyresearch/based-360m").to("cuda", dtype=torch.float16) input...

it works now! 🎉 thank you!