Recommendations for using Managed Service for Kubernetes
Use these recommendations for your PRODUCTION
applications that require:
- High availability and fault tolerance.
- Load scaling.
- Resource isolation.
High availability and fault tolerance
- When creating a cluster, select the Highly available master type. Kubernetes services will be available in the event of a physical server or server rack failure. The Managed Service for Kubernetes Service Level Agreement applies to the configuration with the highly available master.
Load scaling
Use these recommendations if the load on your Managed Service for Kubernetes cluster is constantly growing:
- To make your cluster more robust, create node groups with automatic scaling.
- To reduce the load on the Kubernetes DNS, use NodeLocal DNS. If a cluster is made up of over 50 nodes, use automatic DNS scaling.
- Consider node storage requirements in advance:
- Review disk limits for Compute Cloud.
- Load test your disk subsystem in a test environment.
- To reduce latency at high IOPS, use non-replicated disks.
Isolating resources
Follow these recommendations for applications that use shared Kubernetes cluster resources.
Adjust the values of limits
and requests
for all cluster services:
---
...
containers:
...
resources:
limits:
cpu: 250m
memory: 128Mi
requests:
cpu: 100m
memory: 64Mi
...
Specify vCPU availability in thousandths and RAM in megabytes. The service will not exceed the vCPU or RAM values specified in limits
. Customized requests
enable you to scale cluster nodes using automatic scaling.
To manage pod resources automatically, configure Kubernetes policies:
- Quality of Service for Pods
to create pods in different availability classes. - Limit Ranges
to set limits at the namespace level.