I am currently undertaking the task of auditing EKS Node resource limits, comparing the limits to the requests and actual usage for around 40 applications. I have to pinpoint where resources are being wasted and propose changes to limits/requests for these nodes.
My question for you all is, what percentage above average Usage should I set the resource limits? I know we still need some wiggle room, but say that an application is using on average 531m of Memory, but the limit is at 1000m (1Gb). That limit obviously needs to come down, but where should it come down to? 600m I think would be too close. Is there a rule of thumb to go by here?
Likewise, the same service uses 10.1mcores of CPU on average, but the limit is set to 1core. I know CPU throttling won’t bring down an applicati…
I am currently undertaking the task of auditing EKS Node resource limits, comparing the limits to the requests and actual usage for around 40 applications. I have to pinpoint where resources are being wasted and propose changes to limits/requests for these nodes.
My question for you all is, what percentage above average Usage should I set the resource limits? I know we still need some wiggle room, but say that an application is using on average 531m of Memory, but the limit is at 1000m (1Gb). That limit obviously needs to come down, but where should it come down to? 600m I think would be too close. Is there a rule of thumb to go by here?
Likewise, the same service uses 10.1mcores of CPU on average, but the limit is set to 1core. I know CPU throttling won’t bring down an application, but I’d like to keep wiggle room there to, I’m just not sure how close to bring the limit to the average usage. Any advice?