Traffic splitting is a valuable tool for app development, reducing the risk of outages during app upgrades. With NGINX Service Mesh, it takes less than 10 minutes to implement blue-green and canary deployments. Watch the video demo and check out our step-by-step tutorials.
Welcome to Microservices March!
Microservices March is a month-long virtual festival of microservices activities here at NGINX. Whether you’re already using Kubernetes in production or your interest is just blossoming, you’re sure to find sessions to pique your interest. Check out the schedule on our blog!
Reduce Complexity with Production-Grade Kubernetes
We explain how production-grade Kubernetes solves the challenges of deploying containerized microservices-based apps, which include culture, complexity, and security. In addition to a Kubernetes infrastructure you need a scalable Ingress controller, WAF, and service mesh.
Easy and Robust Single Sign-On with OpenID Connect and NGINX Ingress Controller
NGINX Ingress Controller now supports single sign-on with OpenID Connect. Release 1.10.0 also introduces new configuration queue metrics, annotations on log entries, better validation of annotations and secrets, support for NGINX App Protect user-defined signatures, and more.
Deploying Application Services in Kubernetes, Part 2
We provide guidance on where to deploy application services in a Kubernetes environment, using WAF as an example. Depending on your needs, it can make sense to deploy your WAF at the "front door" of the environment, on the Ingress Controller, per-service, or per-Pod.
Deploying Application Services in Kubernetes, Part 1
We explain why duplicating application services paradoxically can improve overall efficiency: because NetOps and DevOps teams have different mandates, it makes sense for them to select and manage the tools that best suit their specific needs.
Securing Your Apps in Kubernetes with NGINX App Protect
With NGINX Plus Ingress Controller for Kubernetes release 1.8.0, NGINX App Protect can be embedded in the Ingress Controller. This puts WAF protection closer to applications, which is crucial in modern app environments like Kubernetes. It also enables automation and reduces complexity and cost.
Configuring NGINX Plus as an External Load Balancer for Red Hat OCP and Kubernetes
The NGINX Load Balancer Operator is a reference architecture for automating reconfiguration of the external NGINX Plus load balancer for your Red Hat OCP or a Kubernetes cluster, based on changes to the status of the containerized applications.
Announcing NGINX Ingress Controller for Kubernetes Release 1.6.0
Release 1.6.0 of the NGINX Ingress Controller for Kubernetes includes improvements to NGINX Ingress Resources, support for OpenTracing, and much more.
Enabling OpenTracing with NGINX Ingress Controller for Kubernetes
Trace the traffic in your Kubernetes cluster end-to-end with native support for OpenTracing when using the NGINX and NGINXÂ Plus Ingress Controllers for Kubernetes for load balancing. Our step-by-step instructions show you how to get started, using Docker containers and Jaeger.