-
Notifications
You must be signed in to change notification settings - Fork 227
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Use different input size on train and inference #17
Comments
Could you please upload some demo images? So I can figure out the reason. |
From my view, the demo image you uploaded fails in estimating the person 2D location, right? |
It confused me. The network is a fully convolution neural network, why it is failed when changing the input size. Is it a bug in postprocessing? The mesh pose seems right, but the camera pose seems wrong. If the center map prediction failed, we can not get the right human pose since the mesh param is extracted by center map prediction. |
If you change the Coordconv from 128 to 64, the input to the head network will also change. During training, the head network has been familiar with the constant 128-index input from Coordconv. I think the camera parameter is calculated based on the constant 128-index input. I can't make sure whether the camera estimation would be stable if you change its input from 128 to 64. It may affect the feature resolution. After all, the feature is learnt from a different resolution. |
If you try to make full use of the resources, you may mix up 4~8 224x128 image to a 512x512 one. |
We use 256 image input to your trained 512 input model, and adjust the coordconv size, but the result seems not right. Does your network need the same input size on train and test stage?
The text was updated successfully, but these errors were encountered: