Metrics-server CrashLoopBackOff with NEW install by rke

Hello all,
I have installed at least 10 times last one days, but its same every time
Everything runs fine but metrics-server is CrashLoopBackOff

what I understand below section are missing from the pods YAML file and needs to add into Deployment.

I am new to Kubernetes, I have 2 question

  1. I am using rke to install rancher cluster, so why bellow setting are missing from pods to start metric-server?

``
Command:
/metrics-server
–cert-dir=/tmp
–secure-port=4443
–kubelet-preferred-address-types=InternalIP
–kubelet-insecure-tls


2)  what is the best way to add those lines 


### Cluster information:

Kubernetes version:
[rke@rke19-master1 ~]$ kubectl get nodes
NAME           STATUS   ROLES               AGE   VERSION
192.168.0.56   Ready    controlplane,etcd   17m   v1.19.10
192.168.0.57   Ready    controlplane,etcd   17m   v1.19.10
192.168.0.58   Ready    controlplane,etcd   17m   v1.19.10
192.168.0.59   Ready    worker              17m   v1.19.10
192.168.0.60   Ready    worker              17m   v1.19.10
[rke@rke19-master1 ~]$



> [rke@rke19-master1 ~]$ kubectl get pods -n kube-system
> NAME                                       READY   STATUS             RESTARTS   AGE
> calico-kube-controllers-5cc4f54c4f-8wfsq   1/1     Running            0          9m10s
> canal-dxklh                                2/2     Running            0          9m10s
> canal-fgp2j                                2/2     Running            0          9m11s
> canal-fzg2f                                2/2     Running            0          9m11s
> canal-nxgvl                                2/2     Running            0          9m11s
> canal-r59gp                                2/2     Running            0          9m10s
> coredns-67bbc84b8f-2cs8m                   1/1     Running            0          7m6s
> coredns-67bbc84b8f-dbm29                   1/1     Running            0          6m51s
> coredns-autoscaler-fd9d4557c-nkbk9         1/1     Running            0          7m7s
> metrics-server-5b6d79d4f4-ggl57            0/1     CrashLoopBackOff   6          6m46s
> rke-coredns-addon-deploy-job-4rp28         0/1     Completed          0          9m3s
> rke-ingress-controller-deploy-job-p2tll    0/1     Completed          0          6m39s
> rke-metrics-addon-deploy-job-rtf7c         0/1     Completed          0          7m2s
> rke-network-plugin-deploy-job-tkkz5        0/1     Completed          0          9m35s



> [rke@rke19-master1 ~]$ kubectl get pods metrics-server-5b6d79d4f4-ggl57 -n kube-system -o yaml
> apiVersion: v1
> kind: Pod
> metadata:
>   annotations:
>     cni.projectcalico.org/podIP: 10.42.4.3/32
>     cni.projectcalico.org/podIPs: 10.42.4.3/32
>   creationTimestamp: "2021-08-16T23:00:42Z"
>   generateName: metrics-server-5b6d79d4f4-
>   labels:
>     k8s-app: metrics-server
>     pod-template-hash: 5b6d79d4f4
>   managedFields:
>   - apiVersion: v1
>     fieldsType: FieldsV1
>     fieldsV1:
>       f:metadata:
>         f:generateName: {}
>         f:labels:
>           .: {}
>           f:k8s-app: {}
>           f:pod-template-hash: {}
>         f:ownerReferences:
>           .: {}
>           k:{"uid":"fb15b257-4a9d-478b-b461-8b61c165e3db"}:
>             .: {}
>             f:apiVersion: {}
>             f:blockOwnerDeletion: {}
>             f:controller: {}
>             f:kind: {}
>             f:name: {}
>             f:uid: {}
>       f:spec:
>         f:affinity:
>           .: {}
>           f:nodeAffinity:
>             .: {}
>             f:requiredDuringSchedulingIgnoredDuringExecution:
>               .: {}
>               f:nodeSelectorTerms: {}
>         f:containers:
>           k:{"name":"metrics-server"}:
>             .: {}
>             f:args: {}
>             f:image: {}
>             f:imagePullPolicy: {}
>             f:livenessProbe:
>               .: {}
>               f:failureThreshold: {}
>               f:httpGet:
>                 .: {}
>                 f:path: {}
>                 f:port: {}
>                 f:scheme: {}
>               f:periodSeconds: {}
>               f:successThreshold: {}
>               f:timeoutSeconds: {}
>             f:name: {}
>             f:ports:
>               .: {}
>               k:{"containerPort":4443,"protocol":"TCP"}:
>                 .: {}
>                 f:containerPort: {}
>                 f:name: {}
>                 f:protocol: {}
>             f:readinessProbe:
>               .: {}
>               f:failureThreshold: {}
>               f:httpGet:
>                 .: {}
>                 f:path: {}
>                 f:port: {}
>                 f:scheme: {}
>               f:periodSeconds: {}
>               f:successThreshold: {}
>               f:timeoutSeconds: {}
>             f:resources: {}
>             f:securityContext:
>               .: {}
>               f:readOnlyRootFilesystem: {}
>               f:runAsNonRoot: {}
>               f:runAsUser: {}
>             f:terminationMessagePath: {}
>             f:terminationMessagePolicy: {}
>             f:volumeMounts:
>               .: {}
>               k:{"mountPath":"/tmp"}:
>                 .: {}
>                 f:mountPath: {}
>                 f:name: {}
>         f:dnsPolicy: {}
>         f:enableServiceLinks: {}
>         f:priorityClassName: {}
>         f:restartPolicy: {}
>         f:schedulerName: {}
>         f:securityContext: {}
>         f:serviceAccount: {}
>         f:serviceAccountName: {}
>         f:terminationGracePeriodSeconds: {}
>         f:tolerations: {}
>         f:volumes:
>           .: {}
>           k:{"name":"tmp-dir"}:
>             .: {}
>             f:emptyDir: {}
>             f:name: {}
>     manager: kube-controller-manager
>     operation: Update
>     time: "2021-08-16T23:00:42Z"
>   - apiVersion: v1
>     fieldsType: FieldsV1
>     fieldsV1:
>       f:metadata:
>         f:annotations:
>           .: {}
>           f:cni.projectcalico.org/podIP: {}
>           f:cni.projectcalico.org/podIPs: {}
>     manager: calico
>     operation: Update
>     time: "2021-08-16T23:00:47Z"
>   - apiVersion: v1
>     fieldsType: FieldsV1
>     fieldsV1:
>       f:status:
>         f:conditions:
>           k:{"type":"ContainersReady"}:
>             .: {}
>             f:lastProbeTime: {}
>             f:lastTransitionTime: {}
>             f:message: {}
>             f:reason: {}
>             f:status: {}
>             f:type: {}
>           k:{"type":"Initialized"}:
>             .: {}
>             f:lastProbeTime: {}
>             f:lastTransitionTime: {}
>             f:status: {}
>             f:type: {}
>           k:{"type":"Ready"}:
>             .: {}
>             f:lastProbeTime: {}
>             f:lastTransitionTime: {}
>             f:message: {}
>             f:reason: {}
>             f:status: {}
>             f:type: {}
>         f:containerStatuses: {}
>         f:hostIP: {}
>         f:phase: {}
>         f:podIP: {}
>         f:podIPs:
>           .: {}
>           k:{"ip":"10.42.4.3"}:
>             .: {}
>             f:ip: {}
>         f:startTime: {}
>     manager: kubelet
>     operation: Update
>     time: "2021-08-16T23:00:54Z"
>   name: metrics-server-5b6d79d4f4-ggl57
>   namespace: kube-system
>   ownerReferences:
>   - apiVersion: apps/v1
>     blockOwnerDeletion: true
>     controller: true
>     kind: ReplicaSet
>     name: metrics-server-5b6d79d4f4
>     uid: fb15b257-4a9d-478b-b461-8b61c165e3db
>   resourceVersion: "3887"
>   selfLink: /api/v1/namespaces/kube-system/pods/metrics-server-5b6d79d4f4-ggl57
>   uid: af8d4e07-aa3f-4efe-8169-feb37cfd97df
> spec:
>   affinity:
>     nodeAffinity:
>       requiredDuringSchedulingIgnoredDuringExecution:
>         nodeSelectorTerms:
>         - matchExpressions:
>           - key: beta.kubernetes.io/os
>             operator: NotIn
>             values:
>             - windows
>           - key: node-role.kubernetes.io/worker
>             operator: Exists
>   containers:
>   - args:
>     - --cert-dir=/tmp
>     - --secure-port=4443
>     - --kubelet-insecure-tls
>     - --kubelet-preferred-address-types=InternalIP
>     - --logtostderr
>     image: 192.168.0.35:5000/rancher/metrics-server:v0.3.6
>     imagePullPolicy: IfNotPresent
>     livenessProbe:
>       failureThreshold: 3
>       httpGet:
>         path: /livez
>         port: https
>         scheme: HTTPS
>       periodSeconds: 10
>       successThreshold: 1
>       timeoutSeconds: 1
>     name: metrics-server
>     ports:
>     - containerPort: 4443
>       name: https
>       protocol: TCP
>     readinessProbe:
>       failureThreshold: 3
>       httpGet:
>         path: /readyz
>         port: https
>         scheme: HTTPS
>       periodSeconds: 10
>       successThreshold: 1
>       timeoutSeconds: 1
>     resources: {}
>     securityContext:
>       readOnlyRootFilesystem: true
>       runAsNonRoot: true
>       runAsUser: 1000
>     terminationMessagePath: /dev/termination-log
>     terminationMessagePolicy: File
>     volumeMounts:
>     - mountPath: /tmp
>       name: tmp-dir
>     - mountPath: /var/run/secrets/kubernetes.io/serviceaccount
>       name: metrics-server-token-78b6h
>       readOnly: true
>   dnsPolicy: ClusterFirst
>   enableServiceLinks: true
>   nodeName: 192.168.0.59
>   preemptionPolicy: PreemptLowerPriority
>   priority: 2000000000
>   priorityClassName: system-cluster-critical
>   restartPolicy: Always
>   schedulerName: default-scheduler
>   securityContext: {}
>   serviceAccount: metrics-server
>   serviceAccountName: metrics-server
>   terminationGracePeriodSeconds: 30
>   tolerations:
>   - effect: NoExecute
>     operator: Exists
>   - effect: NoSchedule
>     operator: Exists
>   volumes:
>   - emptyDir: {}
>     name: tmp-dir
>   - name: metrics-server-token-78b6h
>     secret:
>       defaultMode: 420
>       secretName: metrics-server-token-78b6h
> status:
>   conditions:
>   - lastProbeTime: null
>     lastTransitionTime: "2021-08-16T23:00:43Z"
>     status: "True"
>     type: Initialized
>   - lastProbeTime: null
>     lastTransitionTime: "2021-08-16T23:00:43Z"
>     message: 'containers with unready status: [metrics-server]'
>     reason: ContainersNotReady
>     status: "False"
>     type: Ready
>   - lastProbeTime: null
>     lastTransitionTime: "2021-08-16T23:00:43Z"
>     message: 'containers with unready status: [metrics-server]'
>     reason: ContainersNotReady
>     status: "False"
>     type: ContainersReady
>   - lastProbeTime: null
>     lastTransitionTime: "2021-08-16T23:00:43Z"
>     status: "True"
>     type: PodScheduled
>   containerStatuses:
>   - containerID: docker://abdfe11278348f3481565a380f5290e4eb7894c6bbcf6373a87ee6bf8d769fd6
>     image: 192.168.0.35:5000/rancher/metrics-server:v0.3.6
>     imageID: docker-pullable://192.168.0.35:5000/rancher/metrics-server@sha256:c9c4e95068b51d6b33a9dccc61875df07dc650abbf4ac1a19d58b4628f89288b
>     lastState:
>       terminated:
>         containerID: docker://abdfe11278348f3481565a380f5290e4eb7894c6bbcf6373a87ee6bf8d769fd6
>         exitCode: 2
>         finishedAt: "2021-08-16T23:15:47Z"
>         reason: Error
>         startedAt: "2021-08-16T23:15:18Z"
>     name: metrics-server
>     ready: false
>     restartCount: 9
>     started: false
>     state:
>       waiting:
>         message: back-off 5m0s restarting failed container=metrics-server pod=metrics-server-5b6d79d4f4-ggl57_kube-system(af8d4e07-aa3f-4efe-8169-feb37cfd97df)
>         reason: CrashLoopBackOff
>   hostIP: 192.168.0.59
>   phase: Running
>   podIP: 10.42.4.3
>   podIPs:
>   - ip: 10.42.4.3
>   qosClass: BestEffort
>   startTime: "2021-08-16T23:00:43Z"


[rke@rke19-master1 ~]$ kubectl logs  metrics-server-5b6d79d4f4-ggl57 -n kube-system
I0816 23:21:20.156497       1 secure_serving.go:116] Serving securely on [::]:4443
[rke@rke19-master1 ~]$

[rke@rke19-master1 ~]$ kubectl get pods metrics-server-5b6d79d4f4-ggl57 -n kube-system -o yaml
apiVersion: v1
kind: Pod
metadata:
  annotations:
    cni.projectcalico.org/podIP: 10.42.4.3/32
    cni.projectcalico.org/podIPs: 10.42.4.3/32
  creationTimestamp: "2021-08-16T23:00:42Z"
  generateName: metrics-server-5b6d79d4f4-
  labels:
    k8s-app: metrics-server
    pod-template-hash: 5b6d79d4f4
  managedFields:
  - apiVersion: v1
    fieldsType: FieldsV1
    fieldsV1:
      f:metadata:
        f:generateName: {}
        f:labels:
          .: {}
          f:k8s-app: {}
          f:pod-template-hash: {}
        f:ownerReferences:
          .: {}
          k:{"uid":"fb15b257-4a9d-478b-b461-8b61c165e3db"}:
            .: {}
            f:apiVersion: {}
            f:blockOwnerDeletion: {}
            f:controller: {}
            f:kind: {}
            f:name: {}
            f:uid: {}
      f:spec:
        f:affinity:
          .: {}
          f:nodeAffinity:
            .: {}
            f:requiredDuringSchedulingIgnoredDuringExecution:
              .: {}
              f:nodeSelectorTerms: {}
        f:containers:
          k:{"name":"metrics-server"}:
            .: {}
            f:args: {}
            f:image: {}
            f:imagePullPolicy: {}
            f:livenessProbe:
              .: {}
              f:failureThreshold: {}
              f:httpGet:
                .: {}
                f:path: {}
                f:port: {}
                f:scheme: {}
              f:periodSeconds: {}
              f:successThreshold: {}
              f:timeoutSeconds: {}
            f:name: {}
            f:ports:
              .: {}
              k:{"containerPort":4443,"protocol":"TCP"}:
                .: {}
                f:containerPort: {}
                f:name: {}
                f:protocol: {}
            f:readinessProbe:
              .: {}
              f:failureThreshold: {}
              f:httpGet:
                .: {}
                f:path: {}
                f:port: {}
                f:scheme: {}
              f:periodSeconds: {}
              f:successThreshold: {}
              f:timeoutSeconds: {}
            f:resources: {}
            f:securityContext:
              .: {}
              f:readOnlyRootFilesystem: {}
              f:runAsNonRoot: {}
              f:runAsUser: {}
            f:terminationMessagePath: {}
            f:terminationMessagePolicy: {}
            f:volumeMounts:
              .: {}
              k:{"mountPath":"/tmp"}:
                .: {}
                f:mountPath: {}
                f:name: {}
        f:dnsPolicy: {}
        f:enableServiceLinks: {}
        f:priorityClassName: {}
        f:restartPolicy: {}
        f:schedulerName: {}
        f:securityContext: {}
        f:serviceAccount: {}
        f:serviceAccountName: {}
        f:terminationGracePeriodSeconds: {}
        f:tolerations: {}
        f:volumes:
          .: {}
          k:{"name":"tmp-dir"}:
            .: {}
            f:emptyDir: {}
            f:name: {}
    manager: kube-controller-manager
    operation: Update
    time: "2021-08-16T23:00:42Z"
  - apiVersion: v1
    fieldsType: FieldsV1
    fieldsV1:
      f:metadata:
        f:annotations:
          .: {}
          f:cni.projectcalico.org/podIP: {}
          f:cni.projectcalico.org/podIPs: {}
    manager: calico
    operation: Update
    time: "2021-08-16T23:00:47Z"
  - apiVersion: v1
    fieldsType: FieldsV1
    fieldsV1:
      f:status:
        f:conditions:
          k:{"type":"ContainersReady"}:
            .: {}
            f:lastProbeTime: {}
            f:lastTransitionTime: {}
            f:message: {}
            f:reason: {}
            f:status: {}
            f:type: {}
          k:{"type":"Initialized"}:
            .: {}
            f:lastProbeTime: {}
            f:lastTransitionTime: {}
            f:status: {}
            f:type: {}
          k:{"type":"Ready"}:
            .: {}
            f:lastProbeTime: {}
            f:lastTransitionTime: {}
            f:message: {}
            f:reason: {}
            f:status: {}
            f:type: {}
        f:containerStatuses: {}
        f:hostIP: {}
        f:phase: {}
        f:podIP: {}
        f:podIPs:
          .: {}
          k:{"ip":"10.42.4.3"}:
            .: {}
            f:ip: {}
        f:startTime: {}
    manager: kubelet
    operation: Update
    time: "2021-08-16T23:00:54Z"
  name: metrics-server-5b6d79d4f4-ggl57
  namespace: kube-system
  ownerReferences:
  - apiVersion: apps/v1
    blockOwnerDeletion: true
    controller: true
    kind: ReplicaSet
    name: metrics-server-5b6d79d4f4
    uid: fb15b257-4a9d-478b-b461-8b61c165e3db
  resourceVersion: "3887"
  selfLink: /api/v1/namespaces/kube-system/pods/metrics-server-5b6d79d4f4-ggl57
  uid: af8d4e07-aa3f-4efe-8169-feb37cfd97df
spec:
  affinity:
    nodeAffinity:
      requiredDuringSchedulingIgnoredDuringExecution:
        nodeSelectorTerms:
        - matchExpressions:
          - key: beta.kubernetes.io/os
            operator: NotIn
            values:
            - windows
          - key: node-role.kubernetes.io/worker
            operator: Exists
  containers:
  - args:
    - --cert-dir=/tmp
    - --secure-port=4443
    - --kubelet-insecure-tls
    - --kubelet-preferred-address-types=InternalIP
    - --logtostderr
    image: 192.168.0.35:5000/rancher/metrics-server:v0.3.6
    imagePullPolicy: IfNotPresent
    livenessProbe:
      failureThreshold: 3
      httpGet:
        path: /livez
        port: https
        scheme: HTTPS
      periodSeconds: 10
      successThreshold: 1
      timeoutSeconds: 1
    name: metrics-server
    ports:
    - containerPort: 4443
      name: https
      protocol: TCP
    readinessProbe:
      failureThreshold: 3
      httpGet:
        path: /readyz
        port: https
        scheme: HTTPS
      periodSeconds: 10
      successThreshold: 1
      timeoutSeconds: 1
    resources: {}
    securityContext:
      readOnlyRootFilesystem: true
      runAsNonRoot: true
      runAsUser: 1000
    terminationMessagePath: /dev/termination-log
    terminationMessagePolicy: File
    volumeMounts:
    - mountPath: /tmp
      name: tmp-dir
    - mountPath: /var/run/secrets/kubernetes.io/serviceaccount
      name: metrics-server-token-78b6h
      readOnly: true
  dnsPolicy: ClusterFirst
  enableServiceLinks: true
  nodeName: 192.168.0.59
  preemptionPolicy: PreemptLowerPriority
  priority: 2000000000
  priorityClassName: system-cluster-critical
  restartPolicy: Always
  schedulerName: default-scheduler
  securityContext: {}
  serviceAccount: metrics-server
  serviceAccountName: metrics-server
  terminationGracePeriodSeconds: 30
  tolerations:
  - effect: NoExecute
    operator: Exists
  - effect: NoSchedule
    operator: Exists
  volumes:
  - emptyDir: {}
    name: tmp-dir
  - name: metrics-server-token-78b6h
    secret:
      defaultMode: 420
      secretName: metrics-server-token-78b6h
status:
  conditions:
  - lastProbeTime: null
    lastTransitionTime: "2021-08-16T23:00:43Z"
    status: "True"
    type: Initialized
  - lastProbeTime: null
    lastTransitionTime: "2021-08-16T23:00:43Z"
    message: 'containers with unready status: [metrics-server]'
    reason: ContainersNotReady
    status: "False"
    type: Ready
  - lastProbeTime: null
    lastTransitionTime: "2021-08-16T23:00:43Z"
    message: 'containers with unready status: [metrics-server]'
    reason: ContainersNotReady
    status: "False"
    type: ContainersReady
  - lastProbeTime: null
    lastTransitionTime: "2021-08-16T23:00:43Z"
    status: "True"
    type: PodScheduled
  containerStatuses:
  - containerID: docker://abdfe11278348f3481565a380f5290e4eb7894c6bbcf6373a87ee6bf8d769fd6
    image: 192.168.0.35:5000/rancher/metrics-server:v0.3.6
    imageID: docker-pullable://192.168.0.35:5000/rancher/metrics-server@sha256:c9c4e95068b51d6b33a9dccc61875df07dc650abbf4ac1a19d58b4628f89288b
    lastState:
      terminated:
        containerID: docker://abdfe11278348f3481565a380f5290e4eb7894c6bbcf6373a87ee6bf8d769fd6
        exitCode: 2
        finishedAt: "2021-08-16T23:15:47Z"
        reason: Error
        startedAt: "2021-08-16T23:15:18Z"
    name: metrics-server
    ready: false
    restartCount: 9
    started: false
    state:
      waiting:
        message: back-off 5m0s restarting failed container=metrics-server pod=metrics-server-5b6d79d4f4-ggl57_kube-system(af8d4e07-aa3f-4efe-8169-feb37cfd97df)
        reason: CrashLoopBackOff
  hostIP: 192.168.0.59
  phase: Running
  podIP: 10.42.4.3
  podIPs:
  - ip: 10.42.4.3
  qosClass: BestEffort
  startTime: "2021-08-16T23:00:43Z"
[rke@rke19-master1 ~]$ ^C
[rke@rke19-master1 ~]$ kubectl logs  metrics-server-5b6d79d4f4-ggl57 -n kube-system
I0816 23:21:20.156497       1 secure_serving.go:116] Serving securely on [::]:4443
[rke@rke19-master1 ~]$ ^C
[rke@rke19-master1 ~]$ kubectl describe  pods metrics-server-5b6d79d4f4-ggl57 -n kube-system
Name:                 metrics-server-5b6d79d4f4-ggl57
Namespace:            kube-system
Priority:             2000000000
Priority Class Name:  system-cluster-critical
Node:                 192.168.0.59/192.168.0.59
Start Time:           Tue, 17 Aug 2021 00:00:43 +0100
Labels:               k8s-app=metrics-server
                      pod-template-hash=5b6d79d4f4
Annotations:          cni.projectcalico.org/podIP: 10.42.4.3/32
                      cni.projectcalico.org/podIPs: 10.42.4.3/32
Status:               Running
IP:                   10.42.4.3
IPs:
  IP:           10.42.4.3
Controlled By:  ReplicaSet/metrics-server-5b6d79d4f4
Containers:
  metrics-server:
    Container ID:  docker://e28b6812965786cd2f520a20dd2adf6cbe9c6a720de905ce16992ed0f4cd7c9e
    Image:         192.168.0.35:5000/rancher/metrics-server:v0.3.6
    Image ID:      docker-pullable://192.168.0.35:5000/rancher/metrics-server@sha256:c9c4e95068b51d6b33a9dccc61875df07dc650abbf4ac1a19d58b4628f89288b
    Port:          4443/TCP
    Host Port:     0/TCP
    Args:
      --cert-dir=/tmp
      --secure-port=4443
      --kubelet-insecure-tls
      --kubelet-preferred-address-types=InternalIP
      --logtostderr
    State:          Waiting
      Reason:       CrashLoopBackOff
    Last State:     Terminated
      Reason:       Error
      Exit Code:    2
      Started:      Tue, 17 Aug 2021 00:21:18 +0100
      Finished:     Tue, 17 Aug 2021 00:21:47 +0100
    Ready:          False
    Restart Count:  11
    Liveness:       http-get https://:https/livez delay=0s timeout=1s period=10s #success=1 #failure=3
    Readiness:      http-get https://:https/readyz delay=0s timeout=1s period=10s #success=1 #failure=3
    Environment:    <none>
    Mounts:
      /tmp from tmp-dir (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from metrics-server-token-78b6h (ro)
Conditions:
  Type              Status
  Initialized       True
  Ready             False
  ContainersReady   False
  PodScheduled      True
Volumes:
  tmp-dir:
    Type:       EmptyDir (a temporary directory that shares a pod's lifetime)
    Medium:
    SizeLimit:  <unset>
  metrics-server-token-78b6h:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  metrics-server-token-78b6h
    Optional:    false
QoS Class:       BestEffort
Node-Selectors:  <none>
Tolerations:     :NoExecuteop=Exists
                 :NoScheduleop=Exists
Events:
  Type     Reason     Age                   From               Message
  ----     ------     ----                  ----               -------
  Normal   Scheduled  21m                   default-scheduler  Successfully assigned kube-system/metrics-server-5b6d79d4f4-ggl57 to 192.168.0.59
  Normal   Pulling    21m                   kubelet            Pulling image "192.168.0.35:5000/rancher/metrics-server:v0.3.6"
  Normal   Pulled     21m                   kubelet            Successfully pulled image "192.168.0.35:5000/rancher/metrics-server:v0.3.6" in 4.687484656s
  Warning  Unhealthy  21m                   kubelet            Readiness probe failed: Get "https://10.42.4.3:4443/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  Unhealthy  21m                   kubelet            Liveness probe failed: Get "https://10.42.4.3:4443/livez": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)
  Warning  Unhealthy  20m                   kubelet            Readiness probe failed: Get "https://10.42.4.3:4443/readyz": dial tcp 10.42.4.3:4443: connect: connection refused
  Warning  Unhealthy  20m (x5 over 21m)     kubelet            Readiness probe failed: HTTP probe failed with statuscode: 404
  Warning  Unhealthy  20m (x5 over 21m)     kubelet            Liveness probe failed: HTTP probe failed with statuscode: 404
  Normal   Killing    20m (x2 over 21m)     kubelet            Container metrics-server failed liveness probe, will be restarted
  Normal   Created    20m (x3 over 21m)     kubelet            Created container metrics-server
  Normal   Started    20m (x3 over 21m)     kubelet            Started container metrics-server
  Warning  BackOff    6m29s (x52 over 18m)  kubelet            Back-off restarting failed container
  Normal   Pulled     83s (x10 over 20m)    kubelet            Container image "192.168.0.35:5000/rancher/metrics-server:v0.3.6" already present on machine
[rke@rke19-master1 ~]$

My first inclination is to get it’s QoS class to guaranteed. This is not like the base deployment of metrics-server that I’m familiar with.

Hi Protosam,
Sorry, I did not understand what you want me to do.
Can you please explain little bit more

You don’t have resource requests set, which affects the QoS for the pod.