walid bettahar
walid bettahar
@kartikayk thanks for the quick reponse !! well ,currently i'm looking for a way to convert the model from .gguf to a native pytorch format , and yes i'm trying...
i want to try with mistral-7b and see if it works well , then i'll figure it out with dolphin-2.2.1-mistral-7b.Q2_K.gguf
thank you for the answers , i have another question sorry for bothering you i only don't get it how to load my own data set for example i have...
atch_size: 4 checkpointer: _component_: torchtune.utils.FullModelHFCheckpointer checkpoint_dir: /tmp/Mistral-7b-v0.1 checkpoint_files: - pytorch_model-00001-of-00002.bin - pytorch_model-00002-of-00002.bin model_type: MISTRAL output_dir: /tmp/Mistral-7b-v0.1 recipe_checkpoint: null compile: false dataset: _component_: torchtune.datasets.instruct_dataset train_on_input: true device: cpu dtype: bf16 enable_activation_checkpointing:...
Actually yes , it's just unstructered text this is why i thought i should be using instruct_dataset # Tokenizer tokenizer: _component_: torchtune.models.mistral.mistral_tokenizer path: /tmp/Mistral-7B-v0.1/tokenizer.model # Dataset dataset: _component_: torchtune.datasets.instruct_dataset train_on_input:...
oh okey thanks , well the data i'm using a text extracted from an html page so it's an unstructured text , so which dataclass i have to use ?
@Jeximo thanks for your answer , i understand that but what i'm trying to do here is to fine-tune my model using a text file similar to this "function1(int ,...