Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Measuring overhead of API calls #179

Open
jtaleric opened this issue Aug 25, 2022 · 2 comments
Open

Measuring overhead of API calls #179

jtaleric opened this issue Aug 25, 2022 · 2 comments

Comments

@jtaleric
Copy link
Contributor

Do we have measurements of the load Cerberus puts on the cluster, since it continuously monitors aspects of the cloud?

Thoughts on using the cluster events to determine go/no-go? Maybe have an option for no sqllite/api calls?

I built a prototype called rlgl (Red Light, Green Light). That simply uses the events to determine if things should stop/go. The only gotcha here is the ttl of events. https://github.com/jtaleric/rlgl -- this won't keep track of things like Cerberus (since it has no db), but for maybe for CI signaling, we can reduce the scope of what we need to know for go/no-go?

@chaitanyaenr
Copy link
Collaborator

We currently do not have the measurements @jtaleric. We will look into measuring the number of API calls every iteration from Cerberus to understand it's overhead. The polling duration is something that's configurable if we do not want to aggressively hit the API.

For the chaos use case, some failures are expected and we query the json stored in the sqlite DB for particular fields and failure counts to pass/fail.

On the extending the ttl of the events will put load on the Etcd AFAIK - something to explore.

@jtaleric
Copy link
Contributor Author

jtaleric commented Sep 6, 2022

On the extending the ttl of the events will put load on the Etcd AFAIK - something to explore.

ack. I am not sure we need to increase the ttl.. It would be good to get some measurements of this.

My .02 here is that if the "bad" events are clearing during reconciliation do we care? It could just be part of the life-cycle.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants