Skip to main content
Version: 3.18 (latest)

Troubleshooting and diagnostics

Logs and diagnostics​

To collect diagnostics, download and install calicoctl somewhere in your $PATH. We recommend installing it as a kubectl plugin by following these directions.

Assuming you installed the binary as a kubectl plugin, you can then create a diagnostics bundle by running the following command:

kubectl calico cluster diags

By default the command collects all logs. Optionally, you can select only those logs that are newer than a relative duration (specified in seconds, minutes or hours). For example:

kubectl calico cluster diags --since=1h

To report a problem, contact Tigera Support.

Alert diagnostics​

The developer console can provide diagnostic information if you have an issue with your alerts and need to contact Support. Please execute the following queries and provide the output along with your alert definition.

GET _watcher/watch/tigera_secure_ee.<cluster_name>.<alert_name>
POST _watcher/watch/tigera_secure_ee.<cluster_name>.<alert_name>/_execute
"action_modes": {
"index_events": "force_simulate"

Check BGP peer status​

If you have connectivity between containers on the same host, and between containers and the Internet, but not between containers on different hosts, it probably indicates a problem in your BGP configuration.

Look at calicoctl node status on each host. It should include output like this:

Calico process is running.

IPv4 BGP status
| | node-to-node mesh | up | 23:30:04 | Established |

IPv6 BGP status
No IPv6 peers found.

Alternatively, you can create a CalicoNodeStatus resource to get BGP session status for the node.

If you do not see this, please check the following.

  • Make sure there is IP connectivity between your hosts.

  • Make sure your network allows the requisite BGP traffic on TCP port 179.

Configure NetworkManager​

Configure NetworkManager before attempting to use Calico Enterprise networking.

NetworkManager manipulates the routing table for interfaces in the default network namespace where Calico Enterprise veth pairs are anchored for connections to containers. This can interfere with the Calico Enterprise agent's ability to route correctly.

The procedure for configuring NetworkManager to ignore Calico Enterprise interfaces varies by Linux distribution. The following steps work best on Ubuntu systems.

  1. Create the following configuration file at /etc/NetworkManager/conf.d/calico.conf.

  2. Restart NetworkManager.

    sudo service network-manager stop
    sudo service network-manager start
  3. Install Calico Enterprise.

  4. Check the interfaces that NetworkManager ignores.

    nmcli dev status

    It should return output indicating that the cali and tunl interfaces are unmanaged.

    If this does not to prevent NetworkManager from interfering with Calico Enterprise networking, try disabling NetworkManager. If disabling NetworkManager does not stop it from interfering with Calico Enterprise networking, you may need to remove NetworkManager. This will require manual network configuration.

Errors when running sudo calicoctl​

If you use sudo for commands, remember that your environment variables are not transferred to the sudo environment. You must run sudo with the -E flag to include your environment variables:

sudo -E calicoctl node diags

or you can set environment variables for sudo commands like this:

sudo DATASTORE_TYPE=kubernetes KUBECONFIG=~/.kube/config calicoctl node run

Also be aware that connection information can be specified as a config file rather than using environment variables. See Installing calicoctl for details.

Error: cnx-node is not ready: BIRD is not ready: BGP not established with​

In most cases, this "unready" status error in Kubernetes means that a particular peer is unreachable in the cluster. Check that BGP connectivity between the two peers is allowed in the environment.

This error can also occur if inactive Node resources are configured for node-to-node mesh. To fix this, decommission the stale nodes.

Linux conntrack table is out of space​

A common problem on Linux systems is running out of space in the conntrack table, which can cause poor iptables performance. This can happen if you run a lot of workloads on a given host, or if your workloads create a lot of TCP connections or bidirectional UDP streams. To avoid this problem, we recommend increasing the conntrack table size using the following commands:

sysctl -w net.netfilter.nf_conntrack_max=1000000
echo "net.netfilter.nf_conntrack_max=1000000" >> /etc/sysctl.conf

Compliance report is not generating at expected time​

By design, reports are scheduled to generate 30 minutes after the specified end time. The reason for this is to allow a certain amount of time to pass for all the relevant data within the specified start and end time to be fully processed and stored. This delay can be modified by setting the TIGERA_COMPLIANCE_JOB_START_DELAY environment variable on the compliance-controller deployment to the desired Golang duration.

GlobalAlert reports error "Trying to create too many buckets"​

"Trying to create too many buckets. Must be less than or equal to: [10000] but was [10001]. This limit can be set by changing the [search.max_buckets] cluster level setting."

The GlobalAlert system has a hard limit of 10000 aggregation keys per query, and will fail to generate alerts if nested aggregations result in the number of keys exceeding this limit. The β€œhealthyβ€œ status of the GlobalAlert will be set to false until the number of aggregation keys returned by the query no longer exceeds this limit.

Careful selection of queries and aggregateBy keys will mitigate this issue. GlobalAlerts should consider the size of the keyspace used in the aggregateBy field and order from least expansive to most. For example: Namespace should precede pod name. Avoid aggregating by source or destination port unless the query selects specific ports. Ephemeral ports used by clients number in the tens of thousands and a single host can trigger this condition.