Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About two softmax outputs #3

Closed
321hallelujah opened this issue Dec 11, 2020 · 2 comments
Closed

About two softmax outputs #3

321hallelujah opened this issue Dec 11, 2020 · 2 comments

Comments

@321hallelujah
Copy link

Hi,
When SSLtraining In your paper, you claim to use two softmax for the pseudo-task. However, in this code, it seems you use tf.split to split the prediction(eg. 8 class for all, and 4 for skip ,the other 4 for transforms), which lead to only one softmax. what is the difference about this, or just my misunderstanding.
Looking forward to your reply.

@sjenni
Copy link
Owner

sjenni commented Dec 12, 2020

Hi,
Thanks for the question. The softmax is applied in the loss function, so in the end, two softmax functions are applied.
The code with splitting should be equivalent to two separate heads. Hope that clarifies the issue? :)
Cheers

@sjenni sjenni closed this as completed Dec 12, 2020
@321hallelujah
Copy link
Author

I got it! the split variable is just fully_connected output of C3D. So it can be split to any pseudo-task. The softmax is applied in every loss function. Finally, the learned temporal-spatial representation is a concatenation style of two pseudo-task.
Is it right? Thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants