Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FAQ: about baseline #3

Open
ahangchen opened this issue May 15, 2018 · 0 comments
Open

FAQ: about baseline #3

ahangchen opened this issue May 15, 2018 · 0 comments

Comments

@ahangchen
Copy link
Owner

Q: (1). In Table 4 of your paper, the last row (TFusion-sup) shows rank-1 accuracy is 73.13%. And my question is:

Your paper adopts DLCE as the supervised learning algorithm, and DLCE achieves rank-1 accuracy 79.51%. Can I say your method degrades performance of supervised learning method, or your method is more suitable for cross datasets scenario? It would be great if you give more details about this.

(Z. Zheng, L. Zheng, and Y. Yang. A discriminatively learned cnn embedding for person re-identification. TOMM, 2017)

A: We implement DLCE in Keras and can't reach 79.51% as they reported, only 75%. Even if we use their MATLAB source code, we can only reach 77% rank-1 accuracy.
In Table4, TFusion-sup rank-1 accuracy is 73.13% because when the vision classifier is very strong, much more powerful than the spatial-temporal model, fusion model will be a little weaker than the vision classifier.
Therefore, our method is more suitable when visual classifier is weak, including cross dataset scenario and some visual-hard scenario like GRID.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant