End of Sale Notice:
Commercial support for NGINX Service Mesh is available to customers who currently have active NGINX Microservices Bundle subscriptions. F5 NGINX announced the End of Sale (EoS) for the NGINX Microservices Bundles as of July 1, 2023.
See our End of Sale announcement for more details.
Release Notes 2.0.0
Release information for F5 NGINX Service Mesh, a configurable, low‑latency infrastructure layer designed to handle a high volume of network‑based interprocess communication among application infrastructure services using application programming interfaces (APIs). Lists of new features and known issues are provided.
NGINX Service Mesh Version 2.0.0
25 April 2023
These release notes provide general information and describe known issues for NGINX Service Mesh version 2.0.0, in the following categories:
Updates
NGINX Service Mesh 2.0.0 includes the following updates:
- NGINX Service Mesh global configuration API has been moved to a Kubernetes Custom Resource Definition. The NGINX Service Mesh API server has been removed. See the API Usage guide for details on how to use the new CRD.
- Removed deprecated auto-injection annotations for Pods in favor of labels.
- Removed deprecated NGINX Ingress Controller annotations for integrating with NGINX Service Mesh in favor of labels.
- Automatic injection is now disabled globally by default, and requires users to opt-in via Namespace or Pod labels. See the Automatic Injection guide for more details.
- Removed
disableAutoInjection
andenabledNamespaces
configuration fields. - Removed deprecated OpenTracing support in favor of OpenTelemetry.
- Fixed issues that would prevent NGINX Service Mesh from deploying in OpenShift.
nginx-mesh-api
component has been renamed tonginx-mesh-controller
.- Helm chart version has been bumped to match the product version.
Resolved Issues
This release includes fixes for the following issues.
-
Pods can’t be created if nginx-mesh-api is unreachable (384)
-
Pods fail to deploy if invalid Jaeger tracing address is set (540)
Known Issues
The following issues are known to be present in this release. Look for updates to these issues in future NGINX Service Mesh release notes.
Lingering invalid RateLimits can cause restart inconsistencies with the NGINX Service Mesh control plane. (658):
The NGINX Service Mesh control plane has a validating webhook that will reject the majority of RateLimits that conflict with an existing RateLimit. However, there are some cases where the validating webhook is unable to determine a conflict. In these cases, the NGINX Service Mesh control plane process will catch the conflict and prevent configuration of the offending RateLimit, but the RateLimit will still be stored in Kubernetes. These RateLimit resources are invalid and can be found by looking for a Warning
event on the RateLimit object. If invalid RateLimits exist and the NGINX Service Mesh control plane restarts, the invalid RateLimits may be configured over the previous valid RateLimits.
Workaround:
When you create a RateLimit resource in Kubernetes, run kubectl describe ratelimit <ratelimit-name>
and check for any Warning
events. If a Warning
event exists, either fix the conflict described in the Warning
event message, or delete the RateLimit by running: kubectl delete ratelimit <ratelimit-name>
.
Duplicate targetPorts in a Service are disregarded (532):
NGINX Service Mesh supports a variety of Service .spec.ports\[]
configurations and honors each port list item with one exception.
If the Service lists multiple port configurations that duplicate .spec.ports\[].targetPort
, the duplicates are disregarded. Only one port configuration is honored for traffic forwarding, authentication, and encryption.
Example invalid configuration:
apiVersion: v1
kind: Service
spec:
ports:
- port: 8080
protocol: TCP
targetPort: 55555
- port: 9090
protocol: TCP
targetPort: 55555
Workaround:
No workaround exists outside of reconfiguring the Service and application. The Service must use unique .spec.ports[].targetPort
values (open up multiple ports on the application workload) or route all traffic to the application workload through the same Service port.
NGINX Service Mesh DNS Suffix support (519):
NGINX Service Mesh only supports the cluster.local
DNS suffix. Services such as Grafana and Prometheus will not work in clusters with a custom DNS suffix.
Workaround:
Ensure your cluster is setup with the default cluster.local
DNS suffix.