-
-
Notifications
You must be signed in to change notification settings - Fork 108
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Openstack exporter hits the default Prometheus scrape timeout #214
Comments
I'd like to echo @peppepetra's suggestion for a caching mechanism to allow for an immediate return of a cached report with tunable metrics collection cycles happening in the background. I suggest in many use cases, Openstack capacity metrics would not impact decision makers if a short, predictable delay between collection and reporting is introduced. |
Well, I don't know about your specific situation but in our case, we discovered that the exporter was scrapping metrics from empy domains created by Heat. |
We are running an OS environment with 3 AZ and a lot of customers (domains) and thousands of assets. After tuning (disabling a lot of metrics and using probe to scrape service [compute,network,volume,...] by service) we still hit timeouts and some scrapes take more than 2 minutes. Would be awesome, someone extends the exporter allowing us to define and use caching? For example: --enable-cache So the result (metrics) got cached for and the next scrape would still return the outdated data but start a fresh metric scan in the background. The outdated data will be displayed until fresh data is available. Would do you think? |
In a big environment with:
openstack-exporter takes around ~30 sec to collect metrics hitting the 15 sec. default scrape timeout.
I have tried disabling slow and deprecated metrics with:
but I am only seeing 1-2 secs of improvements.
It would be nice to have a configurable caching mechanism as described here
Most expensive metrics appear to be:
Disabling the two metrics, the exporter returns in ~15 sec but still hitting the scrape timeout.
The text was updated successfully, but these errors were encountered: