Kubernetes Known Issues
When running in EKS Availability Zone is Unset or Wrong
Symptom
In the WarpStream UI for the cluster you see warpstream-unset-az
set as the availability zone of the agent and/or errors in the agent logs similar to the following:
Context
The WarpStream Agents try to use various methods to determine which availability zone the agent is running in.
When it can't determine the availability zone it falls back to warpstream-unset-az
and logs error messages.
Problem
AWS by default prevents EKS pods from contacting the metadata service to prevent instance metadata leaks. While this is good security practice for normal instances, it prevents services within EKS from querying information about the instance.
Solution
Option A
Use our Helm Chart to deploy WarpStream. The helm chart with it's default configuration will create a Kubernetes ClusterRole
and ClusterRoleBinding
which allows the WarpStream pods to lookup get node they are running on within the Kubernetes API and find the availability zone from node labels.
Option B
Create the appropriate ClusterRole, ClusterRoleBinding, and ServiceAccount so the WarpStream agent can get availability zone information from the Kubernetes API
Then on your WarpStream deployment set the pod service account to warpstream-agent
.
Option C
Modify your EKS Node Launch Template configuration to set http-put-response-hop-limit
to 2.
This will allow the pods running on a EKS instance to connect to the AWS metadata service to find the availability zone.
When running in Kubernetes WarpStream pods end up in the same zone or node
Symptom
Some or all of your WarpStream pods end up running in the same availability zone or on the same Kubernetes node instead of being evenly spread out.
Context
When running workloads in Kubernetes it will try it's best to make sure pods from the same deployment are evenly spread across all nodes and availability zones, however this isn't always possible.
Problem
Depending on Kubernetes cluster configuration and other workloads on the cluster Kubernetes may not evenly deploy WarpStream pods across zones or nodes. Some Kubernetes deployments prioritize bin-packing rather then high availability of workloads. This varies by Kubernetes distribution and is not always configurable.
Solution
Use Kubernetes topologySpreadConstraints
and podAntiAffinity
to force Kubernetes to spread WarpStream pods evenly across zones and nodes. If your WarpStream pods are using our Helm Chart you can set the following in your helm values:
Last updated
Was this helpful?