We provide guidance on where to deploy application services in a Kubernetes environment, using WAF as an example. Depending on your needs, it can make sense to deploy your WAF at the "front door" of the environment, on the Ingress Controller, per-service, or per-Pod.
We explain why duplicating application services paradoxically can improve overall efficiency: because NetOps and DevOps teams have different mandates, it makes sense for them to select and manage the tools that best suit their specific needs.
With NGINX Plus Ingress Controller for Kubernetes release 1.8.0, NGINX App Protect can be embedded in the Ingress Controller. This puts WAF protection closer to applications, which is crucial in modern app environments like Kubernetes. It also enables automation and reduces complexity and cost.
The NGINX Load Balancer Operator is a reference architecture for automating reconfiguration of the external NGINX Plus load balancer for your Red Hat OCP or a Kubernetes cluster, based on changes to the status of the containerized applications.
Release 1.6.0 of the NGINX Ingress Controller for Kubernetes includes improvements to NGINX Ingress Resources, support for OpenTracing, and much more.
Trace the traffic in your Kubernetes cluster end-to-end with native support for OpenTracing when using the NGINX and NGINX Plus Ingress Controllers for Kubernetes for load balancing. Our step-by-step instructions show you how to get started, using Docker containers and Jaeger.
NGINX CMO Rob Whiteley discovers container wars, multi-cloud, serverless, AI/ML, IoT, and a new focus on the enterprise at AWS re:Invent 2017.
Ed Robinson, Chief Marketing Officer, and Owen Garrett, Head of Products, discuss the future of NGINX's open source projects at NGINX Conf 2017.
Use the Ingress controllers for NGINX and NGINX Plus to integrate their sophisticated features into the built-in Kubernetes load-balancing framework.