Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

miscalculation of scores #321

Closed
alaineiturria opened this issue Jul 4, 2018 · 1 comment
Closed

miscalculation of scores #321

alaineiturria opened this issue Jul 4, 2018 · 1 comment

Comments

@alaineiturria
Copy link

To calculate the size of the windows, is it calculated considering the number of anomalies per data set present in the combined_labels.json file or those present in the combined_windows.json file?

I have been able to observe that the scores in the results/null/null_standard_scores.csv file are obtained using the combined_windows.json file, however, it seems to me that the anomalies present in combined_labels.json file are used to calculate the rest of the scores.

@subutai
Copy link
Member

subutai commented Jul 5, 2018

The window size calculation is described in the paper: "we define anomaly window length to be 10% the length of a data file, divided by the number of anomalies in the given file."

For each data file, the number of anomalies and their centers are given in the combined_labels.json file. The resulting windows are then put in combined_windows.json. This whole process is run by the combine_labels.py script.

Some more details are in the NAB whitepaper

For questions like these it might be better to ask on the HTM NAB Forum unless you see a specific bug we need to address.

Hope this helps!

@subutai subutai closed this as completed Jul 5, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants