-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
About replicating monolingual experiments #1
Comments
Here is an update: I have found the Since I do not have access to the
With that script I got the following outputs:
So, these results are closer to the ones reported in Table-5. My question is then: "Is it normal to see such variations from the results in Table-5 or are these different results indicator of a mistake I made?" |
Thanks for your request. I'll look into the issue and let you know asap. |
Your usage and results seem to be fine. I reran the exact configuration and get the same results. The small variation most probably comes from the fact that in our experiments, we cut all vectors to the top 200k thereby increasing efficiency. For test 2, for instance, the term "feuerwaffe" cannot be found in our version. Also note, that in order to keep lists the same lengths, we randomly drop terms from the longer lists. In order to get the exact scores, you might, therefore, need to rerun the experiments multiple times for some languages. If you like to reproduce the exact scores, I can also assist you by forwarding you the exact lists that were used for each individual experiment (but I assume this is not necessary?). |
No need to share the exact list - it is more than enough to know that my usage is correct. Thanks for helping me on this! |
Hi! Thanks for sharing this codebase. I was trying to replicate the results in Table5 of paper.
I have tried 3 languages so far: DE,IT,TR and I couldn't unfortunately replicate the results.
I am probably missing something. Could you help me on this? Let me try to explain what I did for German(DE)
I have downloaded the FastText embedding from here (Specifically, the text version. the file I downloaded is named wiki.de.vec
Then I run the following command after cloning the repository:
Then I checked the automatically created files. For example in w2v_wiki_de_cosine_1_cased.res, I found this:
and in w2v_wiki_de_cosine_2_cased.res:
I was also getting bunch of warning some of which are as follows (written in w2v_wiki_de_cosine_1_cased.out):
Can you tell me what I need to do to run the code successfully? Thanks in advance!
The text was updated successfully, but these errors were encountered: