-
-
Notifications
You must be signed in to change notification settings - Fork 77
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
learning-rate and pretrained model of SAITS #35
Comments
Hi there 👋, Thank you so much for your attention to PyPOTS! If you find PyPOTS helpful to your work, please star⭐️ this repository. Your star is your recognition, which can help more people notice PyPOTS and grow PyPOTS community. It matters and is definitely a kind of contribution to the community. I have received your message and will respond ASAP. Thank you for your patience! 😃 Best, |
Hi, Many thanks for your likes to PyPOTS! Here are my answers: 1). Such a learning rate is tuned by NNI. For all models in the SAITS paper, we use NNI to help tune their hyperparameters to make fair comparisons. I recommend you to read the full paper of SAITS if you'd like to know more details; And if your questions are specifically related to SAITS, you can raise issues in the repo of SAITS. Thank you again for your support! |
Thank you for the answer, I will read the full paper carefully. Sometimes when I found no big val-loss improvement between 100 epochs and 1000 epochs , I may lose the confidence of models, and keen on finding ways to change them. Thank you again for your reply! |
In all experiments I ran, I remember the models usually converge in hundreds of training epochs and it shouldn't take more than 2 hours. There're four GPU cards, but we don't use them for parallelly training a single model, i.e. a model's training time does not get affected by the GPU number. Right, sometimes the loss descends slowly but steadily. You can adjust the learning rate to obtain faster speed. And again, if you want to reproduce the results in the SAITS paper, you'd better use code in https://github.com/WenjieDu/SAITS because there're minor differences in code logic between the repos SAITS and PyPOTS. |
Thanks ! In Table 2 of SAITS paper, the Metrics of MAE / RMSE (like 0.186 / 0.431) are for original data or scaled/normallized data? But the scaled/normallized data is small enough,eg: Thank you very much for the reply! |
I didn't mean RTX5000 can finish all 10K epochs in two hours. When the early stopping strategy is applied, the training procedure doesn't have to finish all 10K epochs. For your 1st question, you can refer to the section For the 2nd one, you can refer to the code in the repo |
Hello, Wenjie,
I tried the PyPOTS with, it awesome! But I have following questions:
(1) During training with SAITS model, I found the learning-rate is recommend to lr = 0.00068277455043675505 in ‘PhysioNet2012_SAITS_best.ini’ file. I am wondering if there are some great methods to get such a learning-rate? (I only know to set 0.001, 0.0001 or such kind of stuffy numbers)
(2) if there are some possible to release the pretrained state_dict .pth file of SAITS(base) and SAITS? Because during training with my custom dataset, I encounter with an early-stop problem inside of 100 epochs, so I decide to see if there will be the same problem with PhysioNet2012 of epochs = 10000.
Or the training log files of SAITS(base) and SAITS would be helpful !
Thank you very much for your reply !
The text was updated successfully, but these errors were encountered: