-
Notifications
You must be signed in to change notification settings - Fork 728
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Guide for using LSTM with PPO2 #231
Comments
Hello, |
Yes, I was using single channel (only 1 image) with the LSTM. |
Sorry, tried and abandoned LSTMs because of lack of success ... |
Hi,
I'm trying to learn navigation policies in a 3D environment while using LSTM as policy for PPO2. I have problem to figure it out the parameters to use.
I usually have an episode that last 200 steps, and I used
n_steps=800
in PPO2+CNN in order to have a quite stable learning. The input of the network was a stack of 4 images.I was wondering how should I change this value while using LSTM. I notices that using 800 means feeding a batch of 800 images to the network and this is a quite long sequence.
Is anyone can give me a suggestion based on your experience?
The text was updated successfully, but these errors were encountered: