bigscience
bigscience copied to clipboard
Is the 13B - unmodified Megatron gpt2 - baseline available? ( tr1-13B-base)
I was super excited to hear about this project! I was wondering if the model is available anywhere?
In the chronicles of tr1-13B-base it says at the end: "All checkpoints converted to HF format and uploaded to HUB.", which I thought meant that it is available on Huggingface, but I can't seem to find it.
Is it available and I'm just not able to find it, or did I misunderstand and it's not available?
The checkpoint for the 1B version is available on HF Hub: https://huggingface.co/bigscience/tr5b-1B3-multilingual-alpha-checkpoints/tree/global_step118500 (Note: You must the select the branch depending on the global step - main branch is empty).