You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to train the semi supervised model on the cardiac ACDC dataset. As I want to obtain the best Dice score as possible, I have set the weight of each loss to zero except the Dice loss which has a weight of 1 (loss with labels).
The problem is that the model does not converge. After a few epochs, there seems to be a collapse and the Dice loss suddenly jumps. The more parameters the model has, the earlier in training this collapse happens (the collapse is almost immediate, before the first epoch with ~20 M parameters but appears after 200 epochs when having only 100 000 parameters). When the collapse happens, the "vxm_dense_flow_loss" or "vxm_dense_flow_resize_loss" (these losses have a 0 weight) explode to a very high value (millions, sometimes a little bit less).
When registering with the learnt weights, the output is an all black image.
Here is what the flow looks like after 600 epochs for the x axis for the smallest model which collapses after 200 epochs:
Values of the flow are between 0 and -250.
Input image values are normalized between 0 and 1.
I have tried to increase the batch size from 1 to 16. I also tried to reduce the learning rate from 1e-4 to 1e-5. It did not solve the issue.
Would you have any idea as to why this happens ?
Thank you for your help.
The text was updated successfully, but these errors were encountered:
Hello.
I am trying to train the semi supervised model on the cardiac ACDC dataset. As I want to obtain the best Dice score as possible, I have set the weight of each loss to zero except the Dice loss which has a weight of 1 (loss with labels).
The problem is that the model does not converge. After a few epochs, there seems to be a collapse and the Dice loss suddenly jumps. The more parameters the model has, the earlier in training this collapse happens (the collapse is almost immediate, before the first epoch with ~20 M parameters but appears after 200 epochs when having only 100 000 parameters). When the collapse happens, the "vxm_dense_flow_loss" or "vxm_dense_flow_resize_loss" (these losses have a 0 weight) explode to a very high value (millions, sometimes a little bit less).
When registering with the learnt weights, the output is an all black image.
Here is what the flow looks like after 600 epochs for the x axis for the smallest model which collapses after 200 epochs:
![flow example voxelmorph](https://private-user-images.githubusercontent.com/50268701/271219390-131fa02a-51bf-4338-ae4f-314ef4e6ea35.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MjE4ODMxODUsIm5iZiI6MTcyMTg4Mjg4NSwicGF0aCI6Ii81MDI2ODcwMS8yNzEyMTkzOTAtMTMxZmEwMmEtNTFiZi00MzM4LWFlNGYtMzE0ZWY0ZTZlYTM1LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNDA3MjUlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjQwNzI1VDA0NDgwNVomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWUxODdmNWE3YjIxMjQ1M2M5YTU4MDI0MmJlMGRmYmYxMWJkZjE1Yjk4Y2Y1Yjc0ZmUxMmRiNGRlYmRiODFjNTMmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0JmFjdG9yX2lkPTAma2V5X2lkPTAmcmVwb19pZD0wIn0.6bEPKly78Ixi7iyCUTbo6rmk6NyplxZoFNkvqzEpYpY)
Values of the flow are between 0 and -250.
Input image values are normalized between 0 and 1.
I have tried to increase the batch size from 1 to 16. I also tried to reduce the learning rate from 1e-4 to 1e-5. It did not solve the issue.
Would you have any idea as to why this happens ?
Thank you for your help.
The text was updated successfully, but these errors were encountered: