Source IP address ¶
By default NGINX uses the content of the header
X-Forwarded-For as the source of truth to get information about the client IP address. This works without issues in L7 if we configure the setting
proxy-real-ip-cidr with the correct information of the IP/network address of trusted external load balancer.
If the ingress controller is running in AWS we need to use the VPC IPv4 CIDR.
Another option is to enable proxy protocol using
In this mode NGINX does not use the content of the header to get the source IP address of the connection.
Proxy Protocol ¶
If you are using a L4 proxy to forward the traffic to the NGINX pods and terminate HTTP/HTTPS there, you will lose the remote endpoint's IP address. To prevent this you could use the Proxy Protocol for forwarding traffic, this will send the connection details before forwarding the actual TCP connection itself.
Support for websockets is provided by NGINX out of the box. No special configuration required.
The only requirement to avoid the close of connections is the increase of the values of
The default value of this settings is
A more adequate value to support websockets is a value higher than one hour (
If the NGINX ingress controller is exposed with a service
type=LoadBalancer make sure the protocol between the loadbalancer and NGINX is TCP.
Optimizing TLS Time To First Byte (TTTFB) ¶
NGINX provides the configuration option ssl_buffer_size to allow the optimization of the TLS record size.
This improves the TLS Time To First Byte (TTTFB).
The default value in the Ingress controller is
4k (NGINX default is
Retries in non-idempotent methods ¶
Since 1.9.13 NGINX will not retry non-idempotent requests (POST, LOCK, PATCH) in case of an error.
The previous behavior can be restored using
retry-non-idempotent=true in the configuration ConfigMap.
- Ingress rules for TLS require the definition of the field
Why endpoints and not services ¶
The NGINX ingress controller does not use Services to route traffic to the pods. Instead it uses the Endpoints API in order to bypass kube-proxy to allow NGINX features like session affinity and custom load balancing algorithms. It also removes some overhead, such as conntrack entries for iptables DNAT.