<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://all.docs.genesys.com/index.php?action=history&amp;feed=atom&amp;title=VM%2FCurrent%2FVMPEGuide%2FVoiceRQServiceMetrics</id>
	<title>VM/Current/VMPEGuide/VoiceRQServiceMetrics - Revision history</title>
	<link rel="self" type="application/atom+xml" href="https://all.docs.genesys.com/index.php?action=history&amp;feed=atom&amp;title=VM%2FCurrent%2FVMPEGuide%2FVoiceRQServiceMetrics"/>
	<link rel="alternate" type="text/html" href="https://all.docs.genesys.com/index.php?title=VM/Current/VMPEGuide/VoiceRQServiceMetrics&amp;action=history"/>
	<updated>2026-04-14T23:47:15Z</updated>
	<subtitle>Revision history for this page on the wiki</subtitle>
	<generator>MediaWiki 1.31.1</generator>
	<entry>
		<id>https://all.docs.genesys.com/index.php?title=VM/Current/VMPEGuide/VoiceRQServiceMetrics&amp;diff=116228&amp;oldid=prev</id>
		<title>Corinneh: Published</title>
		<link rel="alternate" type="text/html" href="https://all.docs.genesys.com/index.php?title=VM/Current/VMPEGuide/VoiceRQServiceMetrics&amp;diff=116228&amp;oldid=prev"/>
		<updated>2022-02-23T20:56:30Z</updated>

		<summary type="html">&lt;p&gt;Published&lt;/p&gt;
&lt;p&gt;&lt;b&gt;New page&lt;/b&gt;&lt;/p&gt;&lt;div&gt;{{ArticlePEServiceMetrics&lt;br /&gt;
|IncludedServiceId=25aca843-1ef0-44f4-9d67-b2a215dcf082&lt;br /&gt;
|CRD=Supports both CRD and annotations&lt;br /&gt;
|Port=12000&lt;br /&gt;
|Endpoint=http://&amp;lt;pod-ipaddress&amp;gt;:12000/metrics&lt;br /&gt;
|MetricsUpdateInterval=30 seconds&lt;br /&gt;
|MetricsDefined=Yes&lt;br /&gt;
|MetricsIntro=You can query Prometheus directly to see all the metrics that the Voice RQ Service exposes. The following metrics are likely to be particularly useful. Genesys does not commit to maintain other currently available Voice RQ Service metrics not documented on this page.&lt;br /&gt;
|PEMetric={{PEMetric&lt;br /&gt;
|Metric=rqnode_clients&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Number of clients connected.&lt;br /&gt;
|UsedFor=Traffic&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_streams&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Number of active streams present.&lt;br /&gt;
|UsedFor=Traffic&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_xreads&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Number of XREAD requests received.&lt;br /&gt;
|UsedFor=Traffic&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_xadds&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Number of XADD requests received.&lt;br /&gt;
|UsedFor=Traffic&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_redis_state&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Current Redis connection state.&lt;br /&gt;
|UsedFor=Errors&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_redis_disconnects&lt;br /&gt;
|Type=counter&lt;br /&gt;
|MetricDescription=The number of Redis disconnects that occurred for the RQ node.&lt;br /&gt;
|UsedFor=Errors&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_consul_leader_error&lt;br /&gt;
|Type=counter&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Number of errors received from Consul during the leadership process.&lt;br /&gt;
|UsedFor=Errors&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_active_master&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Service master role is active.&lt;br /&gt;
|UsedFor=Saturation&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_active_backup&lt;br /&gt;
|Type=gauge&lt;br /&gt;
|Unit=N/A&lt;br /&gt;
|MetricDescription=Service backup role is active.&lt;br /&gt;
|UsedFor=Saturation&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_read_latency&lt;br /&gt;
|Type=histogram&lt;br /&gt;
|Label=le, healthcheck&lt;br /&gt;
|MetricDescription=RQ latency; that is, the time duration between when an event is added to Redis and when it's read via XREAD.&lt;br /&gt;
|UsedFor=Latency&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_add_latency&lt;br /&gt;
|Type=histogram&lt;br /&gt;
|Label=le, healthcheck&lt;br /&gt;
|MetricDescription=RQ latency; that is, the time duration between when a message is received and when it's added to the list.&lt;br /&gt;
|UsedFor=Latency&lt;br /&gt;
}}{{PEMetric&lt;br /&gt;
|Metric=rqnode_redis_latency&lt;br /&gt;
|Type=histogram&lt;br /&gt;
|Label=le&lt;br /&gt;
|MetricDescription=Latency caused by Redis read/write.&lt;br /&gt;
|UsedFor=Latency&lt;br /&gt;
}}&lt;br /&gt;
|AlertsDefined=Yes&lt;br /&gt;
|PEAlert={{PEAlert&lt;br /&gt;
|Alert=Number of Redis streams is too high&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Too many active sessions.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has reached.&lt;br /&gt;
*Check the number of voice, digital, and callback calls in the system.&lt;br /&gt;
|BasedOn=rqnode_streams&lt;br /&gt;
|Threshold=More than 10000 active streams running.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Redis disconnected for 5 minutes&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Redis is not available for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple services, make sure there are no issues with Redis, restart Redis.&lt;br /&gt;
*If the alarm is triggered only for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check to see if there is any issue with the pod.&lt;br /&gt;
|BasedOn=redis_state&lt;br /&gt;
|Threshold=Redis is not available for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Redis disconnected for 10 minutes&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Redis is not available for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple services, make sure there are no issues with Redis, and then restart Redis.&lt;br /&gt;
*If the alarm is triggered only for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check to see if there is any issue with the pod.&lt;br /&gt;
|BasedOn=redis_state&lt;br /&gt;
|Threshold=Redis is not available for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; for 10 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod failed&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; failed.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*One of the containers in the pod has entered a Failed state. Check the Kibana logs for the reason.&lt;br /&gt;
|BasedOn=kube_pod_status_phase&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in Failed state.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod Unknown state&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; in Unknown state.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple services, make sure there are no issues with the Kubernetes cluster.&lt;br /&gt;
*If the alarm is triggered only for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check whether the image is correct and if the container is starting up.&lt;br /&gt;
|BasedOn=kube_pod_status_phase&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; in Unknown state for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod Pending state&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in the Pending state.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If the alarm is triggered for multiple services, make sure the Kubernetes nodes where the pod is running are alive in the cluster.&lt;br /&gt;
*If the alarm is triggered only for the pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;, check the health of the pod.&lt;br /&gt;
|BasedOn=kube_pod_status_phase&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; is in the Pending state for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod not ready for 10 minutes&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; in NotReady state.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*If this alarm is triggered, check whether the CPU is available for the pods.&lt;br /&gt;
*Check whether the port of the pod is running and serving the request.&lt;br /&gt;
|BasedOn=kube_pod_status_ready&lt;br /&gt;
|Threshold=Pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt; in NotReady state for 10 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Container restored repeatedly&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; was repeatedly restarted.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*One of the containers in the pod has entered a failed state. Check the Kibana logs for the reason.&lt;br /&gt;
|BasedOn=kube_pod_container_status_restarts_total&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; was restarted 5 or more times within 15 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod memory greater than 65%&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=High memory usage for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Collect the service logs; raise an investigation ticket.&lt;br /&gt;
|BasedOn=container_memory_working_set_bytes, kube_pod_container_resource_requests_memory_bytes&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; memory usage exceeded 65% for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod memory greater than 80%&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Critical memory usage for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Restart the service.&lt;br /&gt;
*Collect the service logs; raise an investigation ticket.&lt;br /&gt;
|BasedOn=container_memory_working_set_bytes, kube_pod_container_resource_requests_memory_bytes&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; memory usage exceeded 80% for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod CPU greater than 65%&lt;br /&gt;
|Severity=Warning&lt;br /&gt;
|AlertDescription=High CPU load for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Collect the service logs; raise an investigation ticket&lt;br /&gt;
|BasedOn=container_cpu_usage_seconds_total, container_spec_cpu_period&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; CPU usage exceeded 65% for 5 minutes.&lt;br /&gt;
}}{{PEAlert&lt;br /&gt;
|Alert=Pod CPU greater than 80%&lt;br /&gt;
|Severity=Critical&lt;br /&gt;
|AlertDescription=Critical CPU load for pod &amp;lt;nowiki&amp;gt;{{ $labels.pod }}&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
Actions:&lt;br /&gt;
&lt;br /&gt;
*Check whether the horizontal pod autoscaler has triggered and if the maximum number of pods has been reached.&lt;br /&gt;
*Check Grafana for abnormal load.&lt;br /&gt;
*Restart the service.&lt;br /&gt;
*Collect the service logs; raise an investigation ticket.&lt;br /&gt;
|BasedOn=container_cpu_usage_seconds_total, container_spec_cpu_period&lt;br /&gt;
|Threshold=Container &amp;lt;nowiki&amp;gt;{{ $labels.container }}&amp;lt;/nowiki&amp;gt; CPU usage exceeded 80% for 5 minutes.&lt;br /&gt;
}}&lt;br /&gt;
}}&lt;/div&gt;</summary>
		<author><name>Corinneh</name></author>
		
	</entry>
</feed>