We provide guidance on where to deploy application services in a Kubernetes environment, using WAF as an example. Depending on your needs, it can make sense to deploy your WAF at the "front door" of the environment, on the Ingress Controller, per-service, or per-Pod.
We explain why duplicating application services paradoxically can improve overall efficiency: because NetOps and DevOps teams have different mandates, it makes sense for them to select and manage the tools that best suit their specific needs.
The NGINX Load Balancer Operator is a reference architecture for automating reconfiguration of the external NGINX Plus load balancer for your Red Hat OCP or a Kubernetes cluster, based on changes to the status of the containerized applications.
Release 1.6.0 of the NGINX Ingress Controller for Kubernetes includes improvements to NGINX Ingress Resources, support for OpenTracing, and much more.
Trace the traffic in your Kubernetes cluster end-to-end with native support for OpenTracing when using the NGINX and NGINX Plus Ingress Controllers for Kubernetes for load balancing. Our step-by-step instructions show you how to get started, using Docker containers and Jaeger.
"Service mesh" is a hot topic, but as of early 2019 most implementations are not production-ready. In this blog we offer advice on adopting existing technologies that are likely to meet the needs of all but the most complex applications until such time as service mesh is more mature.
NGINX CMO Rob Whiteley discovers container wars, multi-cloud, serverless, AI/ML, IoT, and a new focus on the enterprise at AWS re:Invent 2017.
Ed Robinson, Chief Marketing Officer, and Owen Garrett, Head of Products, discuss the future of open source at NGINX
Use the Ingress controllers for NGINX and NGINX Plus to integrate their sophisticated features into the built-in Kubernetes load-balancing framework.