<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://all.docs.genesys.com/index.php?action=history&amp;feed=atom&amp;title=VM%2FCurrent%2FVMPEGuide%2FVoiceFrontEndServiceMetrics</id>
	<title>VM/Current/VMPEGuide/VoiceFrontEndServiceMetrics - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://all.docs.genesys.com/index.php?action=history&amp;feed=atom&amp;title=VM%2FCurrent%2FVMPEGuide%2FVoiceFrontEndServiceMetrics"/>
	<link rel="alternate" type="text/html" href="https://all.docs.genesys.com/index.php?title=VM/Current/VMPEGuide/VoiceFrontEndServiceMetrics&amp;action=history"/>
	<updated>2026-04-14T23:39:21Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://all.docs.genesys.com/index.php?title=VM/Current/VMPEGuide/VoiceFrontEndServiceMetrics&amp;diff=116230&amp;oldid=prev</id>
		<title>Corinneh: Published</title>
		<link rel="alternate" type="text/html" href="https://all.docs.genesys.com/index.php?title=VM/Current/VMPEGuide/VoiceFrontEndServiceMetrics&amp;diff=116230&amp;oldid=prev"/>
		<updated>2022-02-23T20:56:39Z</updated>

		<summary type="html">&lt;p&gt;Published&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{ArticlePEServiceMetrics&lt;br /&gt;
|IncludedServiceId=fa42b327-7d9a-43c9-b13d-c33ec96146eb&lt;br /&gt;
|CRD=Supports both CRD and annotations&lt;br /&gt;
|Port=9101&lt;br /&gt;
|Endpoint=http://&amp;lt;pod-ipaddress&amp;gt;:9101/metrics&lt;br /&gt;
|MetricsUpdateInterval=30 seconds&lt;br /&gt;
|MetricsDefined=Yes&lt;br /&gt;
|MetricsIntro=Voice FrontEnd Service exposes Genesys-defined, FrontEnd Service–specific metrics as well as some standard Kafka metrics. You can query Prometheus directly to see all the metrics that the FrontEnd Service exposes. The following metrics are likely to be particularly useful. Genesys does not commit to maintain other currently available FrontEnd Service metrics not documented on this page.&lt;br /&gt;
|PEMetric={{PEMetric&lt;br /&gt;
|Metric=kafka_producer_queue_depth&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=kafka_location&lt;br /&gt;
|MetricDescription=Number of Kafka producer pending events.&lt;br /&gt;
|SampleValue=0&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=kafka_producer_queue_age_seconds&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=seconds&lt;br /&gt;
|Label=kafka_location&lt;br /&gt;
|MetricDescription=Age of the oldest producer pending event, in seconds.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=kafka_producer_error_total&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=kafka_location&lt;br /&gt;
|MetricDescription=Number of Kafka producer errors.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=kafka_producer_state&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=kafka_location&lt;br /&gt;
|MetricDescription=Current state of the Kafka producer.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=kafka_producer_biggest_event_size&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Label=kafka_location, topic&lt;br /&gt;
|MetricDescription=Biggest event size so far.&lt;br /&gt;
|SampleValue=515&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=kafka_max_request_size&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Label=kafka_location&lt;br /&gt;
|MetricDescription=Exposed config to compare with biggest event size.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=log_output_bytes_total&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=bytes&lt;br /&gt;
|Label=level, format, module&lt;br /&gt;
|MetricDescription=Total amount of log output, in bytes.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_requests_total&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=tenant&lt;br /&gt;
|MetricDescription=Number of requests.&lt;br /&gt;
|UsedFor=Traffic&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_responses_total&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=tenant&lt;br /&gt;
|MetricDescription=Number of responses for the requests.&lt;br /&gt;
|UsedFor=Traffic&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_sip_nodes_total&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Number of SIP nodes that are alive.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_sip_node_requests_total&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=sip_node_id, tenant&lt;br /&gt;
|MetricDescription=Number of requests to the SIP nodes.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_sip_node_responses_total&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=sip_node_id, tenant, status&lt;br /&gt;
|MetricDescription=Number of responses from the SIP nodes for the requests.&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_sip_node_request_duration_seconds&lt;br /&gt;
|Type=histogram&lt;br /&gt;
|Unit=seconds&lt;br /&gt;
|Label=le, sip_node_id, tenant, status&lt;br /&gt;
|MetricDescription=The duration of time between the SIP node request and the response, measured in seconds.&lt;br /&gt;
|UsedFor=Latency&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=service_version_info&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Label=version&lt;br /&gt;
|MetricDescription=Displays the version of Voice FrontEnd Service that is currently running. In the case of this metric, the labels provide the important information. The metric value is always 1 and does not provide any information.&lt;br /&gt;
|SampleValue=service_version_info{version=&amp;quot;100.0.1000006&amp;quot;} 1&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_health_level&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Health level of the sipfe node:&lt;br /&gt;
&lt;br /&gt;
-1 – fail&amp;lt;br /&amp;gt;&lt;br /&gt;
0 – starting&amp;lt;br /&amp;gt;&lt;br /&gt;
1 – degraded&amp;lt;br /&amp;gt;&lt;br /&gt;
2 – pass&lt;br /&gt;
|SampleValue=2&lt;br /&gt;
|UsedFor=Errors&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=sipfe_health_check_error&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|Label=reason&lt;br /&gt;
|MetricDescription=Health check errors for the sipfe node:&lt;br /&gt;
&lt;br /&gt;
1 – has error&amp;lt;br /&amp;gt;&lt;br /&gt;
0 – no error&lt;br /&gt;
|SampleValue=0&lt;br /&gt;
|UsedFor=Errors&lt;br /&gt;
}}&lt;br /&gt;
|AlertsDefined=Yes&lt;br /&gt;
|PEAlert={{PEAlert&lt;br /&gt;
|Alert=Too many Kafka pending producer events&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Actions:&lt;br /&gt;
&lt;br /&gt;
*Make sure there are no issues with Kafka or &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; pod's CPU and network.&lt;br /&gt;
|BasedOn=kafka_producer_queue_depth&lt;br /&gt;
|Threshold=Too many Kafka producer pending events for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; (more than 100 in 5 minutes).&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Too many received requests without a response&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Actions:&lt;br /&gt;
&lt;br /&gt;
*Collect the service logs for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;; raise an investigation ticket.&lt;br /&gt;
*Restart the service.&lt;br /&gt;
|BasedOn=sipfe_requests_total&lt;br /&gt;
|Threshold=For too many requests, the Front End service at pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; did not send any response (more than 100 requests without a response, measured over 5 minutes).&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=SIP Cluster Service response latency is too high&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple pods, make sure there are no issues with the SIP Cluster Service (CPU, memory, or network overload).&lt;br /&gt;
*If the alarm is triggered only for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check if there is an issue with the pod (CPU, memory, or network overload).&lt;br /&gt;
|BasedOn=sipfe_sip_node_request_duration_seconds_bucket&lt;br /&gt;
|Threshold=Latency for 95% of messages is more than 0.5 seconds for service &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=No requests received&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Absence of received requests for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*For pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, make sure there are no issues with Orchestration Service and Tenant Service or the network to them.&lt;br /&gt;
|BasedOn=sipfe_requests_total&lt;br /&gt;
|Threshold=increase(sipfe_requests_total{pod=~&amp;quot;sipfe-.+&amp;quot;}[5m]) &amp;lt;= 0 and increase(sipfe_requests_total{pod=~&amp;quot;sipfe-.+&amp;quot;}[10m]) &amp;gt; 100&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Too many failure responses sent&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Too many failure responses are sent by the Front End service at pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*For pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, make sure received requests are valid.&lt;br /&gt;
|BasedOn=sipfe_responses_total&lt;br /&gt;
|Threshold=More than 100 failure responses in 5 consecutive minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Too many Kafka producer errors&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Kafka responds with errors at pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*For pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, make sure there are no issues with Kafka.&lt;br /&gt;
|BasedOn=kafka_producer_error_total&lt;br /&gt;
|Threshold=More than 100 errors in 5 consecutive minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Too many SIP Cluster Service error responses&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=SIP Cluster Service responds with errors at pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple pods, make sure there are no issues with the SIP Cluster Service (CPU, memory, or network overload).&lt;br /&gt;
*If the alarm is triggered only for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check if there are issues with requests sent by the pod.&lt;br /&gt;
|BasedOn=sipfe_sip_node_responses_total&lt;br /&gt;
|Threshold=More than 100 errors in 5 consecutive minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Kafka not available&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Kafka is not available for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple services, make sure there are no issues with Kafka, and then restart Kafka.&lt;br /&gt;
*If the alarm is triggered only for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check if there is an issue with the pod.&lt;br /&gt;
|BasedOn=kafka_producer_state&lt;br /&gt;
|Threshold=Kafka is not available for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; for 5 consecutive minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=SIP Node(s) is not available&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=No available SIP Nodes for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple services, make sure there are no issues with SIP Nodes, and then restart SIP Nodes.&lt;br /&gt;
*If the alarm is triggered only for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check if there is an issue with the pod or the network to SIP Nodes.&lt;br /&gt;
|BasedOn=sipfe_sip_nodes_total&lt;br /&gt;
|Threshold=No available SIP Nodes for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; for 5 consecutive minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod status Failed&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Failed state.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Restart the pod. Check to see if there are any issues with the pod after restart.&lt;br /&gt;
|BasedOn=kube_pod_status_phase&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Failed state.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod status Unknown&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Unknown state for 5 minutes.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Restart the pod. Check to see if there are any issues with the pod after restart.&lt;br /&gt;
|BasedOn=kube_pod_status_phase&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Unknown state for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod status Pending&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Pending state for 5 minutes.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Restart the pod. Check to see if there are any issues with the pod after restart.&lt;br /&gt;
|BasedOn=kube_pod_status_phase&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Pending state for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod status NotReady&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in the NotReady state for 5 minutes.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Restart the pod. Check to see if there are any issues with the pod after restart.&lt;br /&gt;
|BasedOn=kube_pod_status_ready&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in the NotReady state for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Container restarted repeatedly&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; was restarted 5 or more times within 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check if a new version of the image was deployed.&lt;br /&gt;
*Check for issues with the Kubernetes cluster.&lt;br /&gt;
|BasedOn=kube_pod_container_status_restarts_total&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; was restarted 5 or more times within 15 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Max replicas is not sufficient for 5 mins&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=For the past 5 minutes, the desired number of replicas is higher than the number of replicas currently available.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check resources available for Kubernetes. Increase resources, if necessary.&lt;br /&gt;
|BasedOn=kube_statefulset_replicas, kube_statefulset_status_replicas&lt;br /&gt;
|Threshold=Desired number of replicas is higher than current available replicas for the past 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pods scaled up greater than 80%&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=For the past 5 minutes, the desired number of replicas is greater than the number of replicas currently available.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check resources available for Kubernetes. Increase resources, if necessary.&lt;br /&gt;
|BasedOn=kube_hpa_status_current_replicas, kube_hpa_spec_max_replicas&lt;br /&gt;
|Threshold=(kube_hpa_status_current_replicas{namespace=&amp;quot;voice&amp;quot;,hpa=&amp;quot;sipfe-node-hpa&amp;quot;} * 100) / kube_hpa_spec_max_replicas{namespace=&amp;quot;voice&amp;quot;,hpa=&amp;quot;sipfe-node-hpa&amp;quot;} &amp;gt; 80  for: 5m&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pods less than Min Replicas&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=The current number of replicas is lower than the minimum number of replicas that should be available. &lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check if Kubernetes cannot deploy new pods or if pods are failing in their status to be active/read.&lt;br /&gt;
|BasedOn=kube_hpa_status_current_replicas, kube_hpa_spec_min_replicas&lt;br /&gt;
|Threshold=For the past 5 minutes, the current number of replicas is lower than the minimum number of replicas that should be available.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod CPU greater than 65%&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=High CPU load for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Collect the service logs for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;; raise an investigation ticket.&lt;br /&gt;
|BasedOn=container_cpu_usage_seconds_total, container_spec_cpu_period&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; CPU usage exceeded 65% for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod CPU greater than 80%&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Critical CPU load for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Restart the service.&lt;br /&gt;
|BasedOn=container_cpu_usage_seconds_total, container_spec_cpu_period&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; CPU usage exceeded 80% for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod memory greater than 65%&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=High memory usage for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Collect the service logs for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;; raise an investigation ticket.&lt;br /&gt;
|BasedOn=container_memory_working_set_bytes, kube_pod_container_resource_requests_memory_bytes&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; memory usage exceeded 65% for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod memory greater than 80%&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Critical memory usage for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Restart the service for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
|BasedOn=container_memory_working_set_bytes, kube_pod_container_resource_requests_memory_bytes&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; memory usage exceeded 80% for 5 minutes.&lt;br /&gt;
}}&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Corinneh</name></author>
		
	</entry>
</feed>