lecifire
lecifire
Hi i encountered the same problem, would need some assistance on how to resolve I tried training on the OLMo1b model. I didn't change anything much in the config yaml...
Hi I am using the official config for Olmo1B. Only thing I've amended was changing the tokenizer to the dolma one. Here are the details of the config. Also wanted...
We started the checkpoint from a run we did and it was produced as an unsharded checkpoint by training.
I also tried resuming the sharded checkpoints produced by training and it had no issues either in resuming or saving subsequently