Skip to main content

Recommended Prometheus metrics

Big picture​

Monitor the Calico Cloud Typha, Felix, and policy component metrics to ensure optimal cluster operation.

Concepts​

Calico Cloud Typha, Felix, and policy components are the most critical to monitor because they are responsible for ensuring networking and security functions are up-to-date and working as expected.

Typha​

Typha is a caching datastore proxy that sits between calico-nodes and Kubernetes API Server. Its primary function is to allow for increased cluster scale by reducing the load on Kubernetes API Server. Without Typha, large clusters (200+ nodes) would need a considerable amount of memory to correspond to the continuous watches and requests from calico-nodes running in the cluster.

Typha maintains a single datastore connection on behalf of all of its clients (processes running in the calico-node pods, with Felix being Typhas’ main client). Typha watches for node, pod, network policy, bgp configuration, and other events on the Kubernetes API Server, caches and deduplicates this data, and fans out these events to its clients.

Felix​

Felix is a component of calico-node and is responsible for Calico Cloud network policy. Felix must be continuously in sync with the datastore to ensure the correct set of policies are applied to the node it is running on.

Typha-felix

About metrics​

Each Calico Cloud component that you want to connect to Prometheus for endpoint metrics has its own configuration (bgp, license, policy, felix, and typha).

Note that Felix is a separate application with metric endpoints, its own core metrics to monitor itself, and a separate port for a second policy metric endpoint.

Metrics​

This section provides metrics recommendations for maintaining optimal cluster operations. Note the following:

  • Threshold values for each metric depend on the cluster size and churn rate.
  • Threshold recommendations are provided where possible, but because each cluster is different, and metrics can depend on cluster churn rate and scale. We recommend that you baseline the cluster to establish numbers that represent normal figures for your cluster.
  • Metrics that start increasing rapidly from the baseline set need attention.
Typha- [Typha general metrics](#typha-general-metrics) - [Typha cluster mesh metrics](#typha-cluster-mesh-metrics) - [Typha client metrics](#typha-client-metrics) - [Typha cache internals](#typha-cache-internals) - [Typha snapshot details](#typha-snapshot-details)Felix- [Policy metrics](#policy-metrics) - [Felix cluster state metrics](#felix-cluster-state-metrics) - [Felix error metrics](#felix-error-metrics) - [Felix time-based metrics](#felix-time-based-metrics)

Typha general metrics​

Datastore cache size​

Datastore cache size
MetricNote: Syncer (type) is Typha's internal name for a client (type).
Individual syncer values:
(typha_cache_size{syncer="bgp"})
(typha_cache_size{syncer="dpi"})
typha_cache_size{syncer="felix"})
(typha_cache_size{syncer="node-status"})
(typha_cache_size{syncer="tunnel-ip-allocation"})

Sum of all syncers:
The sum of all cache sizes (each syncer type has a cache).
sum by (instance) (typha_cache_size)

Largest syncer:
max by (instance) (typha_cache_size)
Example valueExample of: max by (instance) (typha_cache_size{syncer="felix"})

{instance="10.0.1.20:9093"} 661
{instance="10.0.1.31:9093"} 661
ExplanationThe total number of key/value pairs in Typha's in-memory cache.This metric represents the scale of the Calico Cloud datastore as it tracks how many WEPs (pods and services), HEPs (hostendpoints), networksets, globalnetworksets, Calico Cloud Network Policies etc that Typha is aware of across the entire Calico Federation.You can use this metric to monitor individual syncers to Typha (like Felix, BGP etc), or to get a sum of all syncers. We recommend that you monitor the largest syncer but it is completely up to you. This is a good metric to understand how much data is in Typha. Note: If all Typhas are in sync then they should have the same value for this metric.
Threshold value recommendationThe value of this metric will depend on the scale of the Calico Federation and will always increase as WEPs, Calico Cloud network policies and clusters are added. Achieve a baseline first, then monitor for any unexpected increases from the baseline.
Threshold breach symptomsUnexpected increases may indicate memory leaks and performance issues with Typha.
Threshold breach recommendationsCheck CPU usage on Typha pods and Kubernetes nodes. Increase resources if needed, rollout and restart Typha(s) if needed.
Priority levelOptional.

CPU usage​

CPU usage
Metricrate(process_cpu_seconds_total{30s}) * 100
Example value{endpoint="metrics-port", instance="10.0.1.20:9093", job="typha-metrics-svc", namespace="calico-system", pod="calico-typha-6c6cc9fcf7-csbdl", service="typha-metrics-svc"} 0.27999999999999403
ExplanationCPU in use by Typha represented as a percentage of a core.
Threshold value recommendationA spike at startup is normal. It is recommended to achieve a baseline first, then monitor for any unexpected increases from this baseline. A rule of thumb is to investigate maintained CPU usage above 90%.
Threshold breach symptomsUnexpected maintained CPU usage could cause Typha to fall behind in updating its clients (for example, Felix) and could cause delays to policy updates.
Threshold breach recommendationsCheck CPU usage on Kubernetes nodes. If needed, increase resources, and rollout restart Typha(s).
Priority levelRecommended.

Memory usage​

Memory usage
Metricprocess_resident_memory_bytes
Example valueprocess_resident_memory_bytes{endpoint="metrics-port", instance="10.0.1.20:9093", job="typha-metrics-svc", namespace="calico-system", pod="calico-typha-6c6cc9fcf7-csbdl", service="typha-metrics-svc"} 80515072
ExplanationAmount of memory used by Typha.
Threshold value recommendationIt is recommended to achieve a baseline first, then monitor for any unexpected increases from this baseline. A rule of thumb is to investigate if maintained memory usage is above 90% of what is available from the underlying node. The metric can also be used for memory leaks. In this case, the metric would show Typhas' memory consumption rising over time, even though the cluster is in a stable state.
Threshold breach symptomsUnexpected maintained memory usage could cause Typha to fall behind in updating its clients (for example, Felix) and could cause delays to policy updates.
Threshold breach recommendationsCheck memory usage on Kubernetes nodes. Increase resources if needed, and rollout restart Typha(s) if needed.
Priority levelRecommended.

Typha cluster mesh metrics​

The following metrics are applicable only if you have implemented Cluster mesh.

Note that this metric requires a count syntax because you will have a copy of the metric per RemoteClusterConfiguration. As shown in the table, the value 2 = In Sync reflects good connections.

remote_cluster_connection_status\{cluster="foo"\} = 2
remote_cluster_connection_status\{cluster="bar"\} = 2
remote_cluster_connection_status\{cluster="baz"\} = 1

Remote cluster connections (in-sync)​

Remote cluster connections (in-sync)
Metriccount by (instance) (remote_cluster_connection_status == 2)
ExplanationThis represents the number of remote cluster connections that are connected and in sync. Each remote cluster will report a connection_status value from the following list:
- 0 = Not Connected
- 1 = Connecting
- 2 = In Sync
- 3 = Resync in Process
- 4 = Config Change Restart Required

We suggest the count syntax because there will be one copy of remote_cluster_connection_status per cluster: - remote_cluster_connection_status[cluster="foo"] = 2
remote_cluster_connection_status[cluster="bar"] = 2
remote_cluster_connection_status[cluster="baz"] = '

Counting the number of metrics with value 2 returns the number of In Sync clusters.
Threshold value recommendationWhen remote cluster connections are initializing, connection_status values will fluctuate. After the connection is established, this value should be equal to the number of remote clusters in the environment (if everything is in sync).
Threshold breach symptomsN/A
For out-of-sync symptoms, see the out-of-sync metric.
Threshold breach recommendationsN/A
For out-of-sync recommendations, see the out-of-sync metric.
Priority levelRecommended.

Remote cluster connections (out-of-sync)​

The following metrics are applicable only if you have implemented Cluster mesh.

Remote cluster connections (out-of-sync)
Metriccount by (instance) (remote_cluster_connection_status != 2)
ExplanationNumber of remote cluster connections that are not in sync (i.e. resyncing or failing to connect). Each remote cluster will report a connection_status value from the following list:
- 0 = Not Connected
- 1 = Connecting
- 2 = In Sync
- 3 = Resync in Process
- 4 = Config Change Restart Required
Threshold value recommendationThis value should be 2 if everything is in sync. Note: At Typha startup, it is normal to have non-2 values, but it should stabilize at 2 after connections come up.
Threshold breach symptomsTypha will not receive updates from the relevant remote clusters. Connected clients will see stale or partial data from remote clusters.
Threshold breach recommendationsInvestigate Typha's logs where remote cluster connectivity events are logged. Ensure the networking between clusters is not experiencing issues.
Priority levelRecommended.

Typha client metrics​

Total connections accepted​

Total connections accepted
Metrictypha_connections_accepted
Example valuetypha_connections_accepted{endpoint="metrics-port", instance="10.0.1.20:9093", job="typha-metrics-svc", namespace="calico-system", pod="calico-typha-6c6cc9fcf7-csbdl", service="typha-metrics-svc"} 10
ExplanationTotal number of connections accepted over time. This value always increases.
Threshold value recommendationA steady increase over time is normal. Counters rising after a Felix or Typha restart is also normal (as clients get rebalanced). Investigate connection counters that rise rapidly with no Felix or Typha restarts.
Threshold breach symptomsCounters rising when there are no Felix or Typha restarts, or no action that could cause restarts (an upgrade for example), could indicate unexpected Felix or Typha restarts or issues.
Threshold breach recommendationsCheck resource usage on Typha(s) and Kubernetes nodes. Increase resources if needed.
Priority levelOptional.

Client connections actively streaming​

Client connections actively streaming
Metricsum by (instance) (typha_connections_streaming)
Example value{instance="10.0.1.20:9093"} 10
{instance="10.0.1.31:9093"} 5
ExplanationCurrent number of active connections that are "streaming" (have completed the handshake), to this Typha. After a connection has been Accepted (reported in the previous metric), there will be a handshake before the connection is deemed to be actively streaming. This indicates how many clients are connected to a Typha. The sum reflects per-cache metrics as well.
Threshold value recommendationCompare the value for Total Connections Accepted and Client Connections Actively Streaming. The fluctuation of these values should be in-sync with each other if Accepted Connections are turning into Actively Streamed connections. If there is a discrepancy , you should investigate. Note: As always, it is recommended to baseline the relationship between these two metrics to have a sense of what is normal. It is also worth noting that in smaller clusters, it is normal for Typha to be unbalanced. Typha can handle hundreds of connections so it is of no concern if all nodes in a 10-node cluster (for example) connect to the same Typha.
Threshold breach symptomsFelix is not getting updates from Typha. Calico Cloud network policies are out-of-sync.
Threshold breach recommendationsCheck Typha and Felix logs, and rollout restart Typha(s) if needed.
Priority levelRecommended.

Rebalanced client connections​

Rebalanced client connections
Metricrate(typha_connections_dropped{$_rate_interval})
Example value{endpoint="metrics-port", instance="10.0.1.20:9093", job="typha-metrics-svc", namespace="calico-system", pod="calico-typha-6c6cc9fcf7-csbdl", service="typha-metrics-svc"}
ExplanationNumber of client connections dropped to rebalance and share the load across different Typhas.
Threshold value recommendationIt is normal to see this value increasing sometimes. Investigate if connection dropped counters is rising constantly. If all Typhas are dropping connections because all Typhas believe they have too much load, this also warrants investigation.
Threshold breach symptomsDropping connections is rate limited so it should not affect the cluster as a whole. Typha clients, like Felix, will get dropped sometimes (but not constantly), and could result in periodic delays to policy updates.
Threshold breach recommendationsEnsure that the Kubernetes nodes have enough resources.
Priority levelOptional.

99 percentile client fall-behind​

99 percentile client fall-behind
Metricmax by (instance) (typha_client_latency_secs{quantile='0.99'})
Example value{instance="10.0.1.20:9093"} 0.1234
{instance="10.0.1.31:9093"} 0.1234
ExplanationThis metric measures how far behind Typha's client-handling threads are at reading updates.This metric will increase if:
a) The client (e.g Felix) is slow or overloaded and cannot keep up with what Typha is sending or
b) Typha is overloaded and it cannot keep up with writes to all its clients.

This metric is a good indication of your cluster, Felix, and Typha health.
Threshold value recommendationIt is normal for this to spike when new clients connect; they must download and process the snapshot, during which time they will fall slightly behind. Investigate of latency persists.
Threshold breach symptomsTypha clients receiving updates from Typha will be behind in time. Potential symptoms could include Calico Cloud network policies being out-of-sync.
Threshold breach recommendationsCheck Typha and Felix logs and resource usage. It is recommended to focus on Felix logs and resource usage first, as there is generally more overhead with Felix and thus more of a chance of overload. Rollout restart Typha(s) and calico-node(s) if needed.
Priority levelRecommended.

99 percentile client write latency​

99 percentile client write latency
Metricmax by (instance) (typha_client_write_latency_secs)
Example value{instance="10.0.1.20:9093"} 0.007450815
ExplanationTime for Typha to write to a client's socket (for example, Felix).
Threshold value recommendationIf the write latency is increasing, this indicates that a client (for example, Felix) is having an issue, or the network is having an issue. It is normal for intermittent spikes. Investigate any persistent latency.
Threshold breach symptomsTypha clients will lag behind in receiving updates that Typha is sending. Potential symptoms include Calico Cloud network policies being out-of-sync.
Threshold breach recommendationsCheck Felix logs and resource usage.
Priority levelRecommended.

99 percentile client ping latency​

99 percentile client ping latency
Metricmax by (instance) (typha_ping_latency{quantile="0.99"})
Example value{instance="10.0.1.20:9093"} 0.034285331
ExplanationThis metric tracks the round-trip-time from Typha to a client. How long it takes for Typha's clients to respond to pings over the Typha protocol.
Threshold value recommendationAn increase in this metric above 1 second indicates that the clients, network or Typha are more heavily loaded. It is normal for intermittent spikes. Persistent latency above 1 second warrants investigation.
Threshold breach symptomsTypha clients could be behind in time on updates Typha is sending. Potential symptoms include Calico Cloud network policies being out-of-sync.
Threshold breach recommendationsCheck Typha and Felix logs and resource usage. It is recommended to focus on Felix logs and resource usage first, as there is generally more overhead with Felix and thus more of a chance of overload. Check if the node is overloaded and review/increase calico-node/Typha CPU requests if needed. If needed, rollout restart Typha(s) and calico-node(s).
Priority levelRecommended.

Typha cache internals​

99 percentile breadcrumb size​

99 percentile breadcrumb size
Metricmax by (instance) (typha_breadcrumb_size{quantile="0.99"})
ExplanationTypha stores datastore changes as a series of blocks called breadcrumbs. Typha will store updates inside of these breadcrumbs (for example if a pod churned, this would be a single update). Typha can store multiple updates in a single breadcrumb with the default maximum size number being 100.
Threshold value recommendationTypha generating blocks of size 100 during start up is normal. Investigate if Typha is consistently generating blocks of size 90+, which can indicate Typha is overloaded.
Threshold breach symptomsMaintained block of sizes of 100 can indicate that Typha is falling behind on information and updates contained in the datastore. This will lead to Typha clients also falling behind (for example, Calico Cloud network policy object may not be current).
Threshold breach recommendationsCheck Typha logs and resource usage. Check if there is a lot of activity within the cluster that would cause Typha to send large breadcrumbs (for example, a huge amount of pod churn). If possible, reduce churn rate of resources on the cluster.
Priority levelRecommended.

Non-blocking breadcrumbs fraction​

Non-blocking breadcrumb fraction
Metric(sum by (instance) (rate(typha_breadcrumb_non_block{30s})))/((sum by (instance) (rate(typha_breadcrumb_non_block{30s})))+(sum by (instance) (rate(typha_breadcrumb_block{30s}))))
Example value{instance="10.0.1.20:9093"} NaN
ExplanationTypha stores datastore changes as a series of blocks called "breadcrumbs". Each client "follows the breadcrumbs" either by blocking and waiting, or skipping to the next one (non-blocking) if it is already available. Non-blocking breadcrumb actions indicates that Typha is constantly sending breadcrumbs to keep up with the datastore. Blocking breadcrumb actions indicate that Typha and the client have caught up, are up-to-date, and are waiting on the next breadcrumb. This metric will give a ratio between blocking and non-blocking actions that can indicate the health of Typha, its clients, and the cluster.
Threshold value recommendationAs the load on Typha increases, the ratio of skip-ahead, non-blocking reads, increases. If it approaches 100% then Typha may be overloaded (since clients only do non-blocking reads when they're behind).
Threshold breach symptomsConsistent non-blocking breadcrumbs could indicate that Typha is falling behind on information and updates contained in the datastore. This will lead to Typha clients also being behind (for example, Calico Cloud network policy object may not be current).
Threshold breach recommendationsCheck Typha and Felix logs and resource usage. Check if there is a lot of activity within the cluster that would cause Typha to continuously send non-blocking breadcrumbs.
Priority levelRecommended.

Datastore updates total​

Datastore updates total
Metricsum by (instance) (rate(typha_updates_total{30s}))
Example value{instance="10.0.1.20:9093"} 0
ExplanationThe rate of updates from the datastore(s). For example, updates to Pods/Nodes/Policies/etc.
Threshold value recommendationIntermittent spikes are expected. Constant updates indicates a very busy cluster (for example, lots of pod churn).
Threshold breach symptomsConstant updates could lead to overloaded Typhas whereTyphas clients could fall behind.
Threshold breach recommendationsEnsure Typha has enough resources to handle a very dynamic cluster.
Priority levelOptional.

Datastore update skipped (no-ops)​

Datastore update skipped (no-ops)
Metricsum by (instance) (rate(typha_updates_skipped{30s}))
Example value{instance="10.0.1.20:9093"} 0
ExplanationThe number of updates from the datastore that Typha detected were no-ops. For example, an update to a Kubernetes node resource that did not touch any values that is of interest to Calico Cloud. Such updates are not propagated to clients, which saves resources.
Threshold value recommendationN/A
Threshold breach symptomsN/A
Threshold breach recommendationsN/A
Priority levelOptional.

Typha snapshot details​

Snapshot send time​

Median snapshot send time
Metricmax by (instance) (typha_client_snapshot_send_secs{quantile="0.5"})
Example value{instance="10.0.1.20:9093"} NaN
ExplanationThe median time to stream the initial datastore snapshot to each client. It is useful to know the time it takes for a client to receive the data when it connects; it does not include time to process the data.
Threshold value recommendationInvestigate if this value is moving towards 10s of seconds.
Threshold breach symptomsHigh values of this metric could indicate that newly-started clients are taking a long time to get the latest snapshot of the datastore, increasing the window of time where networking/policy updates are not being applied to the dataplane during a restart/upgrade. Typha has a write timeout for writing the snapshot; if a client cannot receive the snapshot within that timeout, it is disconnected. Clients falling behind on information and updates contained in the datastore (for example, Calico Cloud network policy object may not be current).
Threshold breach recommendationsCheck Typha and calico-node logs and resource usage. Check for network congestion. Investigate why a particular calico-node is slow; it is likely on an overloaded node with insufficient CPU).
Priority levelOptional.

Clients requiring grace period​

Clients requiring grace period
Metricsum by (instance) (typha_connections_grace_used)
Example value{instance="10.0.1.20:9093"} 0
ExplanationThe number of Typhas with clients that required a grace period. After sending the snapshot to the client, Typha allows a grace period for the client to catch up to the most recent data. Typha sending the initial snapshot should take < 1 second, but the processing of the snapshot could take longer, so this grace period is there to allow the newly connected client to process the snapshot.
Threshold value recommendationIf this metric is constantly increasing, it can indicate potential performance issues with Typha and clients. It can indicate that performance is being impacted and may warrant investigation.
Threshold breach symptomsHigh values of this metric could indicate clients falling behind on information and updates contained in the datastore (for example, Calico Cloud network policy object may not be current).
Threshold breach recommendationsCheck Typha and calico-node logs and resource usage. Check for network congestion, and determine the root cause.
Priority levelOptional.

Max snapshot size (raw)​

Max snapshot size (raw)
Metricmax(typha_snapshot_raw_bytes)
Example value{} 557359
ExplanationThe raw size in bytes of snapshots sent from Typha to clients.
Threshold value recommendationN/A
Threshold breach symptomsN/A
Threshold breach recommendationsN/A
Priority LevelOptional.

Max snapshot size (compressed)​

Max snapshot size (compressed)
Metricmax(typha_snapshot_compressed_bytes)
Example value{}134845
ExplanationThe compressed size in bytes of snapshots sent from Typha to clients.
Threshold value recommendationThis metric can be helpful for customers to estimate the bandwidth requirements for Felix to startup. For example, if the compressed snapshot size is 20MB in size on average, and 1000 Felix/calico-nodes start up, the bandwidth requirements could be estimated at 20GB between the pool of Typha and the set of Felixes across the network.
Threshold breach symptomsN/A
Threshold breach recommendationsN/A
Priority LevelOptional.

Policy metrics​

note

The following policy metrics are a separate endpoint exposed by Felix that are used in Manager UI. They require special Prometheus configuration to scrape the metrics. For details, see Policy metrics.

Denied traffic​

Denied traffic
Metriccalico_denied_packets
calico_denied_bytes
Example valuecalico_denied_packets{endpoint="calico-metrics-port", instance="ip-10-0-1-30.ca-central-1.compute.internal", job="calico-node-metrics", namespace="calico-system", pod="calico-node-6pcqm", policy="default
ExplanationNumber of packets or bytes that have been dropped by explicit or implicit deny rules. Note that you'll get one instance of calico_denied_packets/bytes for each policy rule that is denying traffic. For example: calico_denied_packets{policy="tier1|fv/policy1|0|deny|-1",scrIP="10.245.13.133"}
Threshold value recommendationThe general rule of thumb is this metric should report zero at a stable state. Any deviation means that policy and traffic have diverged. Achieving a zero state depends on the stability and maturity of your cluster and policy.
Threshold breach symptomsEither unexpected traffic is being denied because of an attack (one example), or expected traffic is being denied because of a misconfiguration in a policy.
Threshold breach recommendationsIf this metric indicates that policy and traffic have diverged, the recommended steps are: Determine if an attack is causing the metric to spike, or if these flows should be allowed. If the flow should indeed be allowed, update the policy or a preceding policy to allow this traffic.
Priority levelRecommended.

Traffic per rule​

Traffic per rule
Metriccnx_policy_rule_bytes
cnx_policy_rule_packets
Example valuecnx_policy_rule_bytes{action="allow", endpoint="calico-metrics-port", instance="ip-10-0-1-20.ca-central-1.compute.internal", job="calico-node-metrics", namespace="calico-system", pod="calico-node-qzpkt", policy="es-kube-controller-access", rule_direction="egress", rule_index="1", service="calico-node-metrics", tier="allow-tigera", traffic_direction="inbound"}
ExplanationNumber of bytes or packets handled by Calico Cloud network policy rules.
Threshold value recommendationThis metric should usually be non-zero (unless expected). A zero value indicates the rule is not matching any packets, and could be surplus to requirements.
Threshold breach symptomsN/A
Threshold breach recommendationsIf this metrics consistently reports a zero value over an acceptable period of time, you can consider removing the policy rule.
Priority LevelOptional.

Connections per policy rule​

Connections per policy rule
Metriccnx_policy_rule_connections
Example valuecnx_policy_rule_connections{endpoint="calico-metrics-port", instance="ip-10-0-1-20.ca-central-1.compute.internal", job="calico-node-metrics", namespace="calico-system", pod="calico-node-qzpkt", policy="es-kube-controller-access", rule_direction="egress", rule_index="0", service="calico-node-metrics", tier="allow-tigera", traffic_direction="outbound"}
ExplanationNumber connections handled by Calico Cloud policy rules.
Threshold value recommendationThis metric is similar to Traffic per Rule but this deals more with flow monitoring. This metric should usually be non-zero. A zero value indicates that the rule is not matching any packets and could be surplus to requirements.
Threshold breach symptomsN/A
Threshold breach recommendationsIf this metrics consistently reports a zero value over an acceptable period of time, this policy rule can be considered for removal.
Priority LevelOptional.

Felix cluster-state metrics​

CPU usage​

CPU usage
Metricrate(process_cpu_seconds_total{30s}) * 100
Example value{endpoint="metrics-port", instance="10.0.1.20:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-qzpkt", service="felix-metrics-svc"}3.1197504199664072
ExplanationCPU in use by calico-node represented as a percentage of a core.
Threshold value recommendationA spike at startup is normal. It is recommended to first achieve a baseline and then monitor for any unexpected increases from this baseline. Investigate if maintained CPU usage goes above 90%.
Threshold breach symptomsUnexpected maintained CPU usage could cause Felix to fall behind and could cause delays to policy updates.
Threshold breach recommendationsCheck CPU usage on Kubernetes nodes. Increase resources if needed, rollout restart calico-node(s) if needed.
Priority levelRecommended.

Memory usage​

Memory usage
Metricprocess_resident_memory_bytes
Example valueprocess_resident_memory_bytes{endpoint="metrics-port", instance="10.0.1.20:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-qzpkt", service="felix-metrics-svc"} 98996224
ExplanationAmount of memory in use by calico-node.
Threshold value recommendationRecommended to achieve a baseline first, then monitor for any unexpected increases from this baseline. Investigate if maintained CPU usage goes above 90% of what is available from the underlying node.
Threshold breach symptomsUnexpected, maintained, memory usage could cause Felix to fall behind and could cause delays to policy updates.
Threshold breach recommendationsCheck memory usage on Kubernetes nodes. Increase resources if needed, rollout restart typha(s) if needed.
Priority levelRecommended.

Active hosts on each endpoint​

Active hosts on each endpoint
Metricfelix_active_local_endpoints
Example valuefelix_active_local_endpoints{endpoint="metrics-port", instance="10.0.1.30:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-6pcqm", service="felix-metrics-svc"} 36
ExplanationNumber of active pod-networked pods, and HEPs, on this node.
Threshold value recommendationThreshold relates to resource limits on the node for example kubelet's max pods setting.
Threshold breach symptomsSuggests Felix is getting out of sync.
Threshold breach recommendationsRolling restart calico-node and report issue to support.
Priority levelOptional.

Active calico nodes​

Active calico nodes
Metricmax(felix_cluster_num_hosts)
Example value{} 3
ExplanationTotal number of nodes in the cluster that have calico-node deployed and running.
Threshold value recommendationThis value should be equal to the number of nodes in the cluster. If there are discrepancies, then calico-nodes on some nodes are having issues.
Threshold breach symptomsCalico Cloud network policies on affected nodes could be out-of-sync.
Threshold breach recommendationsCheck calico-node logs, rollout restart calico-node if needed.
Priority levelRecommended.

Felix cluster policies​

Felix cluster policies
Metricfelix_cluster_num_policies
Example valuefelix_cluster_num_policies{endpoint="metrics-port", instance="10.0.1.20:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-qzpkt", service="felix-metrics-svc"} 58
ExplanationTotal number of Calico Cloud network policies in the cluster.
Threshold value recommendationBecause Calico Cloud is a distributed system, the number of policies should be generally consistent across all nodes. It is expected to have some skew between nodes for a short period of time while they sync, however they should never be out of sync for very long.
Threshold breach symptomsIf nodes are out of sync for long time, calico-nodes may be having issues or experiencing resource contention. Check the Errors Plot to see if there are any iptables errors reported.
Threshold breach recommendationsRedeploy calico-node if issues are seen, and increase resources if needed.
Priority levelOptional.

Felix active local policies​

Felix active local policies
Metricfelix_active_local_policies
Example valuefelix_active_local_policies{endpoint="metrics-port", instance="10.0.1.30:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-6pcqm", service="felix-metrics-svc"} 44
ExplanationTotal number of network policies deployed on per node basis.
Threshold value recommendationThere is no hard limit to active policies. We can handle 1000+ active policies, but it impacts performance, especially if there's pod churn. The best solution is to optimize policies by combining multiple rules into one policy, and make sure that top-level policy selectors are being used.
Threshold breach symptomsN/A
Threshold breach recommendationsRedeploy calico-node if issues are seen, and increase resources if needed.
Priority levelRecommended.

Felix open FDS​

Felix open FDS
Metricsum by (pod) (process_open_fds{pod=~"calico-node.*"})
Example value{pod="calico-node-6pcqm"} 90
ExplanationNumber of opened file descriptors per calico-node pod.
Threshold value recommendationAlert on this metric when it approaches the ulimit (as reported in process_max_fds value). You should not be anywhere near the maximum.
Threshold breach symptomsFelix may become unstable/crash or fail to apply updates as it should. These failures and issues are logged.
Threshold breach recommendationsCheck Felix logs, redeploy calico-node if you see log issues, and increase max_fds value if possible.
Priority LevelOptional.

Felix max FDS​

Felix max FDS
Metricsum by (pod) (process_max_fds{pod=~"calico-node.*"})
Example value{pod="calico-node-qzpkt"} 1048576
ExplanationMaximum number of opened file descriptors allowed per calico-node pod.
Threshold value recommendationN/A
Threshold breach symptomsN/A
Threshold breach recommendationsN/A
Priority levelOptional.

Felix resync started​

Felix resync started
Metricsum(rate(felix_resyncs_started{5m}))
ExplanationThis is the number of times that Typha has reported to Felix that it is re-connecting with the datastore.
Threshold value recommendationOccasional resyncs are normal. Investigate resync counters that rapidly rise.
Threshold breach symptomsTypha pods may be having issues or experiencing resource contention. Some calico-nodes that are paired with Typha pods experiencing issues will not be able to sync with the datastore.
Threshold breach recommendationsInvestigate the root cause to avoid redeploying Typha (which can be very disruptive). Check resource contention and network connectivity from Typha to the datastore to see if Typha is working fine or if the API server is overloaded.
Priority levelRecommended.

Felix dropped logs​

Felix dropped logs
Metricfelix_logs_dropped
Example valuefelix_logs_dropped{endpoint="metrics-port", instance="10.0.1.20:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-qzpkt", service="felix-metrics-svc"} 0
ExplanationThe number of logs Felix has dropped. Note that this metric does not count flow-logs; it counts logs to stdout.
Threshold value recommendationOccasional drops are normal. Investigate if drop counters rapidly rise.
Threshold breach symptomsFelix will drop logs if it cannot keep up with writing them out. These are ordinary code logs, not flow logs. Calico-node may be under resource constraints.
Threshold breach recommendationsCheck CPU usage on calico-nodes and Kubernetes nodes. Increase resources if needed, and rollout restart calico-node(s) if needed.
Priority levelOptional.

Felix error metrics​

IPset errors​

IPset errors
Metricsum(rate(felix_ipset_errors{5m}))
Example value{} 0
ExplanationNumber of ipset creation, modification, and deletion command failures. This metric reports how many times the ipset command has failed when Felix tried to run it. An error can occur when Felix sends bad ipset command data, or the kernel throws an error (potentially because it was too busy to handle this request at that time).
Threshold value recommendationOccasional errors are normal. Investigate error counters that rapidly rise.
Threshold breach symptomsCalico Cloud network policies may not scope all endpoints in network policy rules. Cluster nodes may be under resource contention, which may result in other _error and _seconds metrics rising. Repeated errors could mean some persistent problem (for example, some other process has created an IP set with that name, which is incompatible).
Threshold breach recommendationsSee the Errors Plot graph to determine if the scope is cluster-wide or node-local. Check calico-node logs. Check resource usage and contention on Kubernetes nodes and calico-nodes. Add nodes/resources if needed. If resource contention is not seen, restart calico-node(s) and monitor. Ensure that other process using IPtables are not blocking Calico Cloud network policy management.
Priority levelOptional.

Iptables restore errors​

Iptables restore errors
Metricsum(rate(felix_iptables_restore_errors{5m}))
ExplanationThe number of iptables-restore errors over five minutes. The iptables-restore command is used when Calico Cloud makes a change to iptables. For example, a new WEP or HEP is created, changes to a WEP or HEP or a change to a policy that affects a WEP or HEP.
Threshold value recommendationOccasional errors are normal. Investigate error counters that rapidly rise.
Threshold breach symptomsCalico Cloud network policies are not up to date. Cluster nodes may be under resource contention, which may result in other _error and _seconds metrics rising.
Threshold breach recommendationsSee the Errors Plot graph to determine if the scope is cluster-wide or node-local. Check calico-node logs. Check resource usage and contention on Kubernetes nodes and calico-nodes. Add nodes/resources if needed. If no resource contention is seen, restart calico-node and monitor.
Priority levelOptional.

Iptables save errors​

Iptables save errors
Metricsum(rate(felix_iptables_save_errors{5m}))
Example value{} 0
ExplanationNumber of iptables-save errors. The iptables-save command is run before every iptables-restore command so that Calico Cloud has the current state of iptables.
Threshold value recommendationOccasional errors are normal. Investigate error counters that rapidly rise.
Threshold breach symptomsCalico Cloud network policies are not up to date. Cluster nodes may be under resource contention, which may result in other _error and _seconds metrics rising. Repeated errors could mean some persistent problem (for example, some other process has creating iptables rules that Calico Cloud cannot decode with the version of iptables-save in use).
Threshold breach recommendationsSee the Errors Plot graph to determine if the scope is cluster-wide or node-local. Check calico-node logs. Check resource usage and contention on Kubernetes nodes and calico-nodes. Add nodes/resources if needed. If no resource contention is seen, restart calico-node and monitor.
Priority levelOptional.

Felix log errors​

Felix log errors
Metricsum(rate(felix_log_errors{5m}))
Example value{} 0
ExplanationThe number of times Felix fails to write out a log because the log buffer is full.
Threshold value recommendationOccasional errors are normal. Investigate error counters that rapidly rise.
Threshold breach symptomsCalico-node may be under resource contention, which may result in other _error and _seconds metrics rising.
Threshold breach recommendationsSee the Errors Plot graph to determine if the scope is cluster-wide or node-local. Check resource usage and contention on Kubernetes nodes and calico-nodes. Add nodes/resources if needed. If no resource contention is seen, restart calico-node and monitor.
Priority levelOptional.

Monitor Felix metrics using a graph​

Errors plot graph
Metricrate(felix_ipset_errors{5m}) || rate(felix_iptables_restore_errors[5m]) || rate(felix_iptables_save_errors[5m]) || rate(felix_log_errors{5m})
Example value{endpoint="metrics-port", instance="10.0.1.20:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-qzpkt", service="felix-metrics-svc"} 0
ExplanationChecks if there have been any iptables-save, iptables-restore, or ipset command errors in the past five minutes. Keeps track of what node is reporting which error.
Threshold value recommendationOccasional errors are normal. Investigate error counters that rapidly rise. For this specific metric it is worth focusing on the metric that is spiking, and referencing that metric information.
Threshold breach symptomsDependent on the specific metric that is logging errors.
Threshold breach recommendationsIf more than one metric is rising, check if all rising metrics are related to a specific calico-node. If this is the case, then the issue is local to that calico-node. Check calico-node logs. Check resource usage for the node and calico-node pod. If more than one metric is rising rapidly across all calico-nodes, then it is a cluster-wide issue and cluster health must be checked. Check cluster resource usage, cluster networking/infrastructure health, and restart calico-nodes and calico-typha pods.
Priority levelRecommended.

Felix time-based metrics​

Dataplane apply time quantile 0.5/0.9/0.99​

Dataplane apply time quantile 0.5/0.9/0.99
Metricfelix_int_dataplane_apply_time_seconds{quantile="0.5"}
felix_int_dataplane_apply_time_seconds{quantile="0.9"}
felix_int_dataplane_apply_time_seconds{quantile="0.99"}
Example valuefelix_int_dataplane_apply_time_seconds{quantile="0.5"}:felix_int_dataplane_apply_time_seconds{endpoint="metrics-port", instance="10.0.1.30:9091", job="felix-metrics-svc", namespace="calico-system", pod="calico-node-6pcqm", quantile="0.5", service="felix-metrics-svc"} 0.020859218
ExplanationTime in seconds that it took to apply a dataplane update ,viewed at the median, 90th percentile, and 99th percentile.
Threshold value recommendationThresholds will vary depending on cluster size and rate of churn. It is recommended that a baseline be set to determine a normal threshold value. In the field we have seen >10s in extremely high-scale clusters with 100k+ endpoints and lots of policy/Kubernetes services.
Threshold breach symptomsLarge time-to-apply values will cause a delay between Calico Cloud network policy commits and enforcement in the dataplane. This is dependent on how Calico Cloud waiting for kube-proxy to release the iptables lock, which is influenced by the number of services in use.
Threshold breach recommendationsIncrease cluster resources, and reduce the number of Kubernetes services if possible.
Priority levelRecommended.

Felix route table list seconds quantile 0.5/0.9/0.99​

Felix route table list seconds quantile 0.5/0.9/0.99
Metricfelix_route_table_list_seconds{quantile="0.5"}
felix_route_table_list_seconds{quantile="0.9"}
felix_route_table_list_seconds{quantile="0.99"}
Example valuefelix_route_table_list_seconds{quantile="0.5"}:felix_route_table_list_seconds{endpoint="metrics-port",instance="10.0.1.30:9091",job="felix-metrics-svc",namespace="calico-system", pod="calico-node-6pcqm",quantile="0.5", service="felix-metrics-svc"} 0.000860426
ExplanationTime to list all the interfaces during a resync, viewed at the median, 90th percentile and 99th percentile.
Threshold value recommendationThresholds will vary depending on the number of cali interfaces per node. It is recommended that a baseline be set to determine a normal threshold value.
Threshold breach symptomsHigh values indicate high CPU usage in felix and slow dataplane updates.
Threshold breach recommendationsIncrease cluster resources. Reduce the number of cali interfaces per node where possible.
Priority levelOptional.

Felix graph update time quantile 0.5/0.9/0/99​

Felix graph update time seconds quantile 0.5/0.9/0.99
Metricfelix_calc_graph_update_time_seconds{quantile="0.5"}
felix_calc_graph_update_time_seconds{quantile="0.9"}
felix_calc_graph_update_time_seconds{quantile="0.99"}
Example valuefelix_calc_graph_update_time_seconds{quantile="0.5"}:felix_calc_graph_update_time_seconds{endpoint="metrics-port",instance="10.0.1.30:9091", job="felix-metrics-svc",namespace="calico-system", pod="calico-node-6pcqm",quantile="0.5", service="felix-metrics-svc"} 0.00007129
ExplanationThis metric reports the time taken to update the calculation graph for each datastore on an update call, viewed at the median, 90th percentile and 99th percentile. The calculation graph is the Felix component that takes all the policies/workload endpoints/host endpoints information that it has received from Typha, and distills it down to dataplane updates that are relevant for this node.
Threshold value recommendationAfter start of day (where we will typically get a large update), then values should be sub 1 second (with occasional blips to 1+ seconds). Should be measured in milliseconds with the occasional blip to a second or two. Investigate if the result is constantly in values of seconds.
Threshold breach symptomsHigh values indicate high CPU usage in felix and slow dataplane updates.
Threshold breach recommendationsIncrease cluster resources. Check calico-node logs. Rollout restart calico-node(s) if needed.
Priority levelRecommended.