-
Bug
-
Resolution: Cannot Reproduce
-
Normal
-
None
-
4.15
-
Quality / Stability / Reliability
-
False
-
-
None
-
Important
-
No
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
Description of problem:
In Reliability (loaded longrun, the load is stable) test, the 'console-operator' container's cpu memory increased.
Version-Release number of selected component (if applicable):
4.15.0-0.nightly-2023-12-09-012410
How reproducible:
I didn't see this issue in 4.14 testing
Steps to Reproduce:
1. Install a AWS cluster with 3 masters, 3 workers. vm_type: m5.xlarge 2. Run reliability-v2 test https://github.com/openshift/svt/tree/master/reliability-v2. The test will long run and simulate multiple customers usage on the cluster. config: 1 admin, 10 dev-test, 5 dev-prod, 1 dev-cron(10cronjobs)
Actual results:
console-operator container's cpu and memory increased during the test
Expected results:
console-operator container's cpu and memory usage should be stable
Additional info:
% oc get po -n openshift-console-operator
NAME READY STATUS RESTARTS AGE
console-operator-5f8ff488c7-ccnz5 2/2 Running 6 (11d ago) 11d
% oc describe po console-operator-5f8ff488c7-ccnz5 -n openshift-console-operator
Name: console-operator-5f8ff488c7-ccnz5
Namespace: openshift-console-operator
Priority: 2000000000
Priority Class Name: system-cluster-critical
Service Account: console-operator
Node: ip-10-0-67-9.us-east-2.compute.internal/10.0.67.9
Start Time: Mon, 11 Dec 2023 09:49:15 +0800
Labels: name=console-operator
pod-template-hash=5f8ff488c7
Annotations: k8s.ovn.org/pod-networks:
{"default":{"ip_addresses":["10.130.0.41/23"],"mac_address":"0a:58:0a:82:00:29","gateway_ips":["10.130.0.1"],"routes":[{"dest":"10.128.0.0...
k8s.v1.cni.cncf.io/network-status:
[{
"name": "ovn-kubernetes",
"interface": "eth0",
"ips": [
"10.130.0.41"
],
"mac": "0a:58:0a:82:00:29",
"default": true,
"dns": {}
}]
openshift.io/scc: restricted-v2
seccomp.security.alpha.kubernetes.io/pod: runtime/default
Status: Running
SeccompProfile: RuntimeDefault
IP: 10.130.0.41
IPs:
IP: 10.130.0.41
Controlled By: ReplicaSet/console-operator-5f8ff488c7
Containers:
console-operator:
Container ID: cri-o://808f374466c03357a42433fb496378996180a694d2ee6b99f5431e028fda8834
Image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10d91d35b86a547ca0e1b4f561a94a39b66af642fc0e701028ce3dca7b5f5966
Image ID: c03dd739e4449beb429b76a880636bad94c4085ef7e60f818edc1d9cd594ea78
Port: 60000/TCP
Host Port: 0/TCP
Command:
console
operator
Args:
-v=2
--config=/var/run/configmaps/config/controller-config.yaml
State: Running
Started: Mon, 11 Dec 2023 09:50:02 +0800
Last State: Terminated
Reason: Error
Message: W1211 01:49:39.641371 1 cmd.go:234] Using insecure, self-signed certificates
F1211 01:49:39.641644 1 cmd.go:159] mkdir /tmp/serving-cert-3011261756: read-only file system
Exit Code: 255
Started: Mon, 11 Dec 2023 09:49:39 +0800
Finished: Mon, 11 Dec 2023 09:49:39 +0800
Ready: True
Restart Count: 3
Requests:
cpu: 10m
memory: 100Mi
Liveness: http-get https://:8443/healthz delay=0s timeout=1s period=10s #success=1 #failure=3
Readiness: http-get https://:8443/readyz delay=0s timeout=1s period=10s #success=1 #failure=3
Environment:
CONSOLE_IMAGE: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3445cf238b78f6a603c595218f092df6f7eb68f96e5293e11e11821b51e68eb
DOWNLOADS_IMAGE: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8ae7550b0f6905d47acfd2529718d697a85de55a9ee8a256591361693f931bea
RELEASE_VERSION: 4.15.0-0.nightly-2023-12-09-012410
OPERATOR_NAME: console-operator
POD_NAME: console-operator-5f8ff488c7-ccnz5 (v1:metadata.name)
Mounts:
/var/run/configmaps/config from config (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-zwssw (ro)
/var/run/secrets/serving-cert from serving-cert (rw)
conversion-webhook-server:
Container ID: cri-o://28638bc6dbfa3fed0e4983ef1f94105015240d91357e6a7fbef8e8169213338c
Image: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10d91d35b86a547ca0e1b4f561a94a39b66af642fc0e701028ce3dca7b5f5966
Image ID: c03dd739e4449beb429b76a880636bad94c4085ef7e60f818edc1d9cd594ea78
Port: 9443/TCP
Host Port: 0/TCP
Command:
console
crdconvert
Args:
--port=9443
--tls-cert-file=/var/run/secrets/webhook-serving-cert/tls.crt
--tls-private-key-file=/var/run/secrets/webhook-serving-cert/tls.key
State: Running
Started: Mon, 11 Dec 2023 09:50:02 +0800
Last State: Terminated
Reason: Error
Message: F1211 01:49:39.820093 1 config.go:27] open /var/run/secrets/webhook-serving-cert/tls.crt: no such file or directory
Exit Code: 255
Started: Mon, 11 Dec 2023 09:49:39 +0800
Finished: Mon, 11 Dec 2023 09:49:39 +0800
Ready: True
Restart Count: 3
Requests:
cpu: 10m
memory: 100Mi
Liveness: http-get https://:9443/readyz delay=0s timeout=1s period=10s #success=1 #failure=3
Readiness: http-get https://:9443/readyz delay=0s timeout=1s period=10s #success=1 #failure=3
Environment: <none>
Mounts:
/var/run/configmaps/config from config (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-zwssw (ro)
/var/run/secrets/webhook-serving-cert from webhook-serving-cert (rw)
Conditions:
Type Status
Initialized True
Ready True
ContainersReady True
PodScheduled True
Volumes:
config:
Type: ConfigMap (a volume populated by a ConfigMap)
Name: console-operator-config
Optional: false
serving-cert:
Type: Secret (a volume populated by a Secret)
SecretName: serving-cert
Optional: true
webhook-serving-cert:
Type: Secret (a volume populated by a Secret)
SecretName: webhook-serving-cert
Optional: true
kube-api-access-zwssw:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
ConfigMapName: openshift-service-ca.crt
ConfigMapOptional: <nil>
QoS Class: Burstable
Node-Selectors: node-role.kubernetes.io/master=
Tolerations: node-role.kubernetes.io/master:NoSchedule op=Exists
node.kubernetes.io/memory-pressure:NoSchedule op=Exists
node.kubernetes.io/not-ready:NoExecute op=Exists for 120s
node.kubernetes.io/unreachable:NoExecute op=Exists for 120s
Events: <none>