Time-LLM icon indicating copy to clipboard operation
Time-LLM copied to clipboard

has anyone run it on kaggle?

Open m6129 opened this issue 1 year ago • 5 comments

Hi. Is it possible to run and test this model on Kaggle?

m6129 avatar Apr 05 '24 08:04 m6129

Hi. Is it possible to run and test this model on Kaggle?

Yes, without any doubt, our model can be used for Kaggle competitions. However, one consideration that might need to be taken into account is the time consumed for training and inference due to the volume of data. It's possible to try lightweight modes such as using BERT and GPT-2 based on Time-LLM.

kwuking avatar Apr 16 '24 04:04 kwuking

Hi. Is it possible to run and test this model on Kaggle?

Yes, without any doubt, our model can be used for Kaggle competitions. However, one consideration that might need to be taken into account is the time consumed for training and inference due to the volume of data. It's possible to try lightweight modes such as using BERT and GPT-2 based on Time-LLM.

There is a problem with the environment in kaggle. https://www.kaggle.com/code/mrantonzaitsev/time-llm/notebook

Perhaps you have some ideas or examples how the problem could be solved? Anyway thanks for your hard work.

m6129 avatar Apr 16 '24 07:04 m6129

updating requirements.txt didn't help, still can't launch on kaggle

m6129 avatar May 07 '24 08:05 m6129

https://www.kaggle.com/code/mrantonzaitsev/time-llm/notebook

Hi, I have checked the error in the URL you provided. The error log is as follows:

RuntimeError: CUDA error: invalid device ordinal CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.

This issue usually results from a discrepancy in the number of CUDA devices or their IDs. The script you executed is configured to use 8 GPUs. Do you have a matching number of GPUs available in your current Kaggle environment? It is advised that you run it using 8 A100 GPUs for optimal results. If you do not have enough GPUs, you might consider reducing the batch size, or using GPT-2/Bert as a base model replacement (this functionality has been fully implemented in Time-LLM). Additionally, it's important to ensure that the num_workers parameter in your script matches the number of GPUs. Once again, thank you for your interest in our work.

kwuking avatar May 17 '24 13:05 kwuking

Kaggle may always provide one GPU, so just try to delete "--multi_gpu" in scripts.

zachysun avatar May 24 '24 01:05 zachysun