Web Server Load Balancing with NGINX Plus

Traffic splitting – dividing traffic between different versions of an app running in the same environment – is a valuable tool for app development because it helps reduce the risk of bad customer experiences when a new app or version is released. In Kubernetes environments, the Ingress controller is often the default choice for splitting traffic, but a service mesh is a good alternative in more complex environments, where it enables you to compare versions of individual microservices. For example, you might want to do a canary deployment behind your mobile frontend, between two different versions of your geo‑location microservices API.

Setting up traffic splitting with NGINX Service Mesh is a simple task which you can complete in less than 10 minutes. In this demo, we show how to use NGINX Service Mesh to implement both a blue‑green and a canary deployment of a new application version.

To help you easily implement these rollout strategies in your own NGINX Service Mesh environment, we also provide step-by-step tutorials for blue‑green and canary deployments.

For more info on traffic splitting and other traffic management patterns, read our blog How to Improve Resilience in Kubernetes with Advanced Traffic Management on our blog.

Also check out these video demos of NGINX Service Mesh features:

Get Started with NGINX Service Mesh Today

NGINX Service Mesh is completely free and available for immediate download and can be deployed in less than 10 minutes! To get started, check out our docs and let us know how it goes via GitHub.

Are You Service Mesh Ready?

Cover image

Explore the factors that determine whether you need a service mesh and which is best for your environment

About The Author

Kate Osborn

Software Engineer

About F5 NGINX

F5, Inc. is the company behind NGINX, the popular open source project. We offer a suite of technologies for developing and delivering modern applications. Together with F5, our combined solution bridges the gap between NetOps and DevOps, with multi-cloud application services that span from code to customer.

Learn more at or join the conversation by following @nginx on Twitter.