There are several clusters accessed by a couple of applications, regularly, sometimes simultaneously.
The applications use the Kube API implementation for Java.
Too often, the Kube API server is prevented from doing its job, as it thinks too many requests have been received by the aforementioned applications.
What would be the best way to avoid this problem, without implementing some complicated request coordination for the applications using the Kubernetes API?
All I’ve found is information on how to limit requests further, etc.
However, I do not want to limit requests, I want the Kube API server to accept the ones received, no matter how much. Ideally it should just queue all requests, until they are all fulfilled in a safe non-overloading manner.