TheGullahanMaster
TheGullahanMaster
You supply the path to the checkpoint via `--ckpt`, ie. `--ckpt samples/000001.pt`. Also don't forget to supply original hyperparams, like image size, style dim, if we use SN etc...
Also BLOOM seems to work fine for unlimited generation, with only a few adjustments to remove the generation character limit, though i had to modify the "config" to be compatible...
In utils.py, i edited the "build_gpt2_config like this: "`def build_gpt2_config( vocab_size: int = 10000, bos_token_id: int = 0, eos_token_id: int = 0, max_length: int = 2048, dropout: float = 0.0,...
and commented out this in aitextgen.py ` if prompt: prompt_num_tokens = list(prompt_tensors["input_ids"].shape)[1] #assert prompt_num_tokens < model_max_length( # self.model.config #), f"The prompt is too large for the model. ({prompt_num_tokens} tokens)"` for...