Added new affinity mode for maximum session stickyness. Fixes kubernetes/ingress-nginx#4475
This commit is contained in:
parent
9bcc386823
commit
8c63ef856e
16 changed files with 541 additions and 55 deletions
|
@ -9,6 +9,7 @@ Session affinity can be configured using the following annotations:
|
||||||
|Name|Description|Value|
|
|Name|Description|Value|
|
||||||
| --- | --- | --- |
|
| --- | --- | --- |
|
||||||
|nginx.ingress.kubernetes.io/affinity|Type of the affinity, set this to `cookie` to enable session affinity|string (NGINX only supports `cookie`)|
|
|nginx.ingress.kubernetes.io/affinity|Type of the affinity, set this to `cookie` to enable session affinity|string (NGINX only supports `cookie`)|
|
||||||
|
|nginx.ingress.kubernetes.io/affinity-mode|The affinity mode defines how sticky a session is. Use `balanced` to redistribute some sessions when scaling pods or `persistent` for maximum stickyness.|`balanced` (default) or `persistent`|
|
||||||
|nginx.ingress.kubernetes.io/session-cookie-name|Name of the cookie that will be created|string (defaults to `INGRESSCOOKIE`)|
|
|nginx.ingress.kubernetes.io/session-cookie-name|Name of the cookie that will be created|string (defaults to `INGRESSCOOKIE`)|
|
||||||
|nginx.ingress.kubernetes.io/session-cookie-path|Path that will be set on the cookie (required if your [Ingress paths][ingress-paths] use regular expressions)|string (defaults to the currently [matched path][ingress-paths])|
|
|nginx.ingress.kubernetes.io/session-cookie-path|Path that will be set on the cookie (required if your [Ingress paths][ingress-paths] use regular expressions)|string (defaults to the currently [matched path][ingress-paths])|
|
||||||
|nginx.ingress.kubernetes.io/session-cookie-max-age|Time until the cookie expires, corresponds to the `Max-Age` cookie directive|number of seconds|
|
|nginx.ingress.kubernetes.io/session-cookie-max-age|Time until the cookie expires, corresponds to the `Max-Age` cookie directive|number of seconds|
|
||||||
|
|
|
@ -17,6 +17,7 @@ You can add these Kubernetes annotations to specific Ingress objects to customiz
|
||||||
|---------------------------|------|
|
|---------------------------|------|
|
||||||
|[nginx.ingress.kubernetes.io/app-root](#rewrite)|string|
|
|[nginx.ingress.kubernetes.io/app-root](#rewrite)|string|
|
||||||
|[nginx.ingress.kubernetes.io/affinity](#session-affinity)|cookie|
|
|[nginx.ingress.kubernetes.io/affinity](#session-affinity)|cookie|
|
||||||
|
|[nginx.ingress.kubernetes.io/affinity-mode](#session-affinity)|"balanced" or "persistent"|
|
||||||
|[nginx.ingress.kubernetes.io/auth-realm](#authentication)|string|
|
|[nginx.ingress.kubernetes.io/auth-realm](#authentication)|string|
|
||||||
|[nginx.ingress.kubernetes.io/auth-secret](#authentication)|string|
|
|[nginx.ingress.kubernetes.io/auth-secret](#authentication)|string|
|
||||||
|[nginx.ingress.kubernetes.io/auth-type](#authentication)|basic or digest|
|
|[nginx.ingress.kubernetes.io/auth-type](#authentication)|basic or digest|
|
||||||
|
@ -151,6 +152,8 @@ If the Application Root is exposed in a different path and needs to be redirecte
|
||||||
The annotation `nginx.ingress.kubernetes.io/affinity` enables and sets the affinity type in all Upstreams of an Ingress. This way, a request will always be directed to the same upstream server.
|
The annotation `nginx.ingress.kubernetes.io/affinity` enables and sets the affinity type in all Upstreams of an Ingress. This way, a request will always be directed to the same upstream server.
|
||||||
The only affinity type available for NGINX is `cookie`.
|
The only affinity type available for NGINX is `cookie`.
|
||||||
|
|
||||||
|
The annotation `nginx.ingress.kubernetes.io/affinity-mode` defines the stickyness of a session. Setting this to `balanced` (default) will redistribute some sessions if a deployment gets scaled up, therefore rebalancing the load on the servers. Setting this to `persistent` will not rebalance sessions to new servers, therefore providing maximum stickyness.
|
||||||
|
|
||||||
!!! attention
|
!!! attention
|
||||||
If more than one Ingress is defined for a host and at least one Ingress uses `nginx.ingress.kubernetes.io/affinity: cookie`, then only paths on the Ingress using `nginx.ingress.kubernetes.io/affinity` will use session cookie affinity. All paths defined on other Ingresses for the host will be load balanced through the random selection of a backend server.
|
If more than one Ingress is defined for a host and at least one Ingress uses `nginx.ingress.kubernetes.io/affinity: cookie`, then only paths on the Ingress using `nginx.ingress.kubernetes.io/affinity` will use session cookie affinity. All paths defined on other Ingresses for the host will be load balanced through the random selection of a backend server.
|
||||||
|
|
||||||
|
|
|
@ -199,11 +199,12 @@ func TestAffinitySession(t *testing.T) {
|
||||||
fooAnns := []struct {
|
fooAnns := []struct {
|
||||||
annotations map[string]string
|
annotations map[string]string
|
||||||
affinitytype string
|
affinitytype string
|
||||||
|
affinitymode string
|
||||||
name string
|
name string
|
||||||
}{
|
}{
|
||||||
{map[string]string{annotationAffinityType: "cookie", annotationAffinityCookieName: "route"}, "cookie", "route"},
|
{map[string]string{annotationAffinityType: "cookie", annotationAffinityMode: "balanced", annotationAffinityCookieName: "route"}, "cookie", "balanced", "route"},
|
||||||
{map[string]string{annotationAffinityType: "cookie", annotationAffinityCookieName: "route1"}, "cookie", "route1"},
|
{map[string]string{annotationAffinityType: "cookie", annotationAffinityMode: "persistent", annotationAffinityCookieName: "route1"}, "cookie", "persistent", "route1"},
|
||||||
{map[string]string{annotationAffinityType: "cookie", annotationAffinityCookieName: ""}, "cookie", "INGRESSCOOKIE"},
|
{map[string]string{annotationAffinityType: "cookie", annotationAffinityMode: "balanced", annotationAffinityCookieName: ""}, "cookie", "balanced", "INGRESSCOOKIE"},
|
||||||
{map[string]string{}, "", ""},
|
{map[string]string{}, "", ""},
|
||||||
{nil, "", ""},
|
{nil, "", ""},
|
||||||
}
|
}
|
||||||
|
@ -213,6 +214,10 @@ func TestAffinitySession(t *testing.T) {
|
||||||
r := ec.Extract(ing).SessionAffinity
|
r := ec.Extract(ing).SessionAffinity
|
||||||
t.Logf("Testing pass %v %v", foo.affinitytype, foo.name)
|
t.Logf("Testing pass %v %v", foo.affinitytype, foo.name)
|
||||||
|
|
||||||
|
if (r.Mode != foo.affinitymode) {
|
||||||
|
t.Errorf("Returned %v but expected %v for Name", r.Mode, foo.affinitymode)
|
||||||
|
}
|
||||||
|
|
||||||
if r.Cookie.Name != foo.name {
|
if r.Cookie.Name != foo.name {
|
||||||
t.Errorf("Returned %v but expected %v for Name", r.Cookie.Name, foo.name)
|
t.Errorf("Returned %v but expected %v for Name", r.Cookie.Name, foo.name)
|
||||||
}
|
}
|
||||||
|
|
|
@ -28,6 +28,7 @@ import (
|
||||||
|
|
||||||
const (
|
const (
|
||||||
annotationAffinityType = "affinity"
|
annotationAffinityType = "affinity"
|
||||||
|
annotationAffinityMode = "affinity-mode"
|
||||||
// If a cookie with this name exists,
|
// If a cookie with this name exists,
|
||||||
// its value is used as an index into the list of available backends.
|
// its value is used as an index into the list of available backends.
|
||||||
annotationAffinityCookieName = "session-cookie-name"
|
annotationAffinityCookieName = "session-cookie-name"
|
||||||
|
@ -57,6 +58,8 @@ var (
|
||||||
type Config struct {
|
type Config struct {
|
||||||
// The type of affinity that will be used
|
// The type of affinity that will be used
|
||||||
Type string `json:"type"`
|
Type string `json:"type"`
|
||||||
|
// The affinity mode, i.e. how sticky a session is
|
||||||
|
Mode string `json:"mode"`
|
||||||
Cookie
|
Cookie
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -136,6 +139,12 @@ func (a affinity) Parse(ing *networking.Ingress) (interface{}, error) {
|
||||||
at = ""
|
at = ""
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Check the afinity mode that will be used
|
||||||
|
am, err := parser.GetStringAnnotation(annotationAffinityMode, ing)
|
||||||
|
if err != nil {
|
||||||
|
am = ""
|
||||||
|
}
|
||||||
|
|
||||||
switch at {
|
switch at {
|
||||||
case "cookie":
|
case "cookie":
|
||||||
cookie = a.cookieAffinityParse(ing)
|
cookie = a.cookieAffinityParse(ing)
|
||||||
|
@ -146,6 +155,7 @@ func (a affinity) Parse(ing *networking.Ingress) (interface{}, error) {
|
||||||
|
|
||||||
return &Config{
|
return &Config{
|
||||||
Type: at,
|
Type: at,
|
||||||
|
Mode: am,
|
||||||
Cookie: *cookie,
|
Cookie: *cookie,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
|
@ -67,6 +67,7 @@ func TestIngressAffinityCookieConfig(t *testing.T) {
|
||||||
|
|
||||||
data := map[string]string{}
|
data := map[string]string{}
|
||||||
data[parser.GetAnnotationWithPrefix(annotationAffinityType)] = "cookie"
|
data[parser.GetAnnotationWithPrefix(annotationAffinityType)] = "cookie"
|
||||||
|
data[parser.GetAnnotationWithPrefix(annotationAffinityMode)] = "balanced"
|
||||||
data[parser.GetAnnotationWithPrefix(annotationAffinityCookieName)] = "INGRESSCOOKIE"
|
data[parser.GetAnnotationWithPrefix(annotationAffinityCookieName)] = "INGRESSCOOKIE"
|
||||||
data[parser.GetAnnotationWithPrefix(annotationAffinityCookieExpires)] = "4500"
|
data[parser.GetAnnotationWithPrefix(annotationAffinityCookieExpires)] = "4500"
|
||||||
data[parser.GetAnnotationWithPrefix(annotationAffinityCookieMaxAge)] = "3000"
|
data[parser.GetAnnotationWithPrefix(annotationAffinityCookieMaxAge)] = "3000"
|
||||||
|
@ -84,6 +85,10 @@ func TestIngressAffinityCookieConfig(t *testing.T) {
|
||||||
t.Errorf("expected cookie as affinity but returned %v", nginxAffinity.Type)
|
t.Errorf("expected cookie as affinity but returned %v", nginxAffinity.Type)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if nginxAffinity.Mode != "balanced" {
|
||||||
|
t.Errorf("expected balanced as affinity mode but returned %v", nginxAffinity.Mode)
|
||||||
|
}
|
||||||
|
|
||||||
if nginxAffinity.Cookie.Name != "INGRESSCOOKIE" {
|
if nginxAffinity.Cookie.Name != "INGRESSCOOKIE" {
|
||||||
t.Errorf("expected INGRESSCOOKIE as session-cookie-name but returned %v", nginxAffinity.Cookie.Name)
|
t.Errorf("expected INGRESSCOOKIE as session-cookie-name but returned %v", nginxAffinity.Cookie.Name)
|
||||||
}
|
}
|
||||||
|
|
|
@ -568,6 +568,10 @@ func (n *NGINXController) getBackendServers(ingresses []*ingress.Ingress) ([]*in
|
||||||
ups.SessionAffinity.AffinityType = anns.SessionAffinity.Type
|
ups.SessionAffinity.AffinityType = anns.SessionAffinity.Type
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if ups.SessionAffinity.AffinityMode == "" {
|
||||||
|
ups.SessionAffinity.AffinityMode = anns.SessionAffinity.Mode
|
||||||
|
}
|
||||||
|
|
||||||
if anns.SessionAffinity.Type == "cookie" {
|
if anns.SessionAffinity.Type == "cookie" {
|
||||||
cookiePath := anns.SessionAffinity.Cookie.Path
|
cookiePath := anns.SessionAffinity.Cookie.Path
|
||||||
if anns.Rewrite.UseRegex && cookiePath == "" {
|
if anns.Rewrite.UseRegex && cookiePath == "" {
|
||||||
|
|
|
@ -141,6 +141,7 @@ func (s Backend) HashInclude(field string, v interface{}) (bool, error) {
|
||||||
// +k8s:deepcopy-gen=true
|
// +k8s:deepcopy-gen=true
|
||||||
type SessionAffinityConfig struct {
|
type SessionAffinityConfig struct {
|
||||||
AffinityType string `json:"name"`
|
AffinityType string `json:"name"`
|
||||||
|
AffinityMode string `json:"mode"`
|
||||||
CookieSessionAffinity CookieSessionAffinity `json:"cookieSessionAffinity"`
|
CookieSessionAffinity CookieSessionAffinity `json:"cookieSessionAffinity"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -152,6 +152,9 @@ func (sac1 *SessionAffinityConfig) Equal(sac2 *SessionAffinityConfig) bool {
|
||||||
if sac1.AffinityType != sac2.AffinityType {
|
if sac1.AffinityType != sac2.AffinityType {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
if sac1.AffinityMode != sac2.AffinityMode {
|
||||||
|
return false
|
||||||
|
}
|
||||||
if !(&sac1.CookieSessionAffinity).Equal(&sac2.CookieSessionAffinity) {
|
if !(&sac1.CookieSessionAffinity).Equal(&sac2.CookieSessionAffinity) {
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
57
rootfs/etc/nginx/lua/affinity/balanced.lua
Normal file
57
rootfs/etc/nginx/lua/affinity/balanced.lua
Normal file
|
@ -0,0 +1,57 @@
|
||||||
|
-- An affinity mode which makes sure connections are rebalanced when a deployment is scaled.
|
||||||
|
-- The advantage of this mode is that the load on the pods will be redistributed.
|
||||||
|
-- The drawback of this mode is that, when scaling up a deployment, roughly (n-c)/n users
|
||||||
|
-- will lose their session, where c is the current number of pods and n is the new number of
|
||||||
|
-- pods.
|
||||||
|
--
|
||||||
|
-- This class extends/implements the abstract class balancer.sticky.
|
||||||
|
--
|
||||||
|
local math = require("math")
|
||||||
|
local resty_chash = require("resty.chash")
|
||||||
|
local util = require("util")
|
||||||
|
|
||||||
|
local _M = {}
|
||||||
|
|
||||||
|
-- Consider the situation of N upstreams one of which is failing.
|
||||||
|
-- Then the probability to obtain failing upstream after M iterations would be close to (1/N)**M.
|
||||||
|
-- For the worst case (2 upstreams; 20 iterations) it would be ~10**(-6)
|
||||||
|
-- which is much better then ~10**(-3) for 10 iterations.
|
||||||
|
local MAX_UPSTREAM_CHECKS_COUNT = 20
|
||||||
|
|
||||||
|
local function get_routing_key(self)
|
||||||
|
return self:get_cookie(), nil
|
||||||
|
end
|
||||||
|
|
||||||
|
local function set_routing_key(self, key)
|
||||||
|
self:set_cookie(key)
|
||||||
|
end
|
||||||
|
|
||||||
|
local function pick_new_upstream(self, failed_upstreams)
|
||||||
|
for i = 1, MAX_UPSTREAM_CHECKS_COUNT do
|
||||||
|
local key = string.format("%s.%s.%s", ngx.now() + i, ngx.worker.pid(), math.random(999999))
|
||||||
|
|
||||||
|
local new_upstream = self.instance:find(key)
|
||||||
|
|
||||||
|
if not failed_upstreams[new_upstream] then
|
||||||
|
return new_upstream, key
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
return nil, nil
|
||||||
|
end
|
||||||
|
|
||||||
|
function _M.new(self, sticky_balancer, backend)
|
||||||
|
local o = sticky_balancer or {}
|
||||||
|
|
||||||
|
local nodes = util.get_nodes(backend.endpoints)
|
||||||
|
|
||||||
|
-- override sticky.balancer methods
|
||||||
|
o.instance = resty_chash:new(nodes)
|
||||||
|
o.get_routing_key = get_routing_key
|
||||||
|
o.set_routing_key = set_routing_key
|
||||||
|
o.pick_new_upstream = pick_new_upstream
|
||||||
|
|
||||||
|
return sticky_balancer
|
||||||
|
end
|
||||||
|
|
||||||
|
return _M
|
53
rootfs/etc/nginx/lua/affinity/persistent.lua
Normal file
53
rootfs/etc/nginx/lua/affinity/persistent.lua
Normal file
|
@ -0,0 +1,53 @@
|
||||||
|
-- An affinity mode which makes sure a session is always routed to the same endpoint.
|
||||||
|
-- The advantage of this mode is that a user will never lose his session.
|
||||||
|
-- The drawback of this mode is that when scaling up a deployment, sessions will not
|
||||||
|
-- be rebalanced.
|
||||||
|
--
|
||||||
|
local util = require("util")
|
||||||
|
local util_nodemap = require("util.nodemap")
|
||||||
|
|
||||||
|
local _M = {}
|
||||||
|
|
||||||
|
local function get_routing_key(self)
|
||||||
|
local cookie_value = self:get_cookie()
|
||||||
|
|
||||||
|
if cookie_value then
|
||||||
|
-- format <timestamp>.<workder-pid>.<routing-key>
|
||||||
|
local routing_key = string.match(cookie_value, '[^\\.]+$')
|
||||||
|
|
||||||
|
if routing_key == nil then
|
||||||
|
local err = string.format("Failed to extract routing key from cookie '%s'!", cookie_value)
|
||||||
|
return nil, err
|
||||||
|
end
|
||||||
|
|
||||||
|
return routing_key, nil
|
||||||
|
end
|
||||||
|
|
||||||
|
return nil, nil
|
||||||
|
end
|
||||||
|
|
||||||
|
local function set_routing_key(self, key)
|
||||||
|
local value = string.format("%s.%s.%s", ngx.now(), ngx.worker.pid(), key)
|
||||||
|
self:set_cookie(value);
|
||||||
|
end
|
||||||
|
|
||||||
|
local function pick_new_upstream(self, failed_upstreams)
|
||||||
|
return self.instance:random_except(failed_upstreams)
|
||||||
|
end
|
||||||
|
|
||||||
|
function _M.new(self, sticky_balancer, backend)
|
||||||
|
local o = sticky_balancer or {}
|
||||||
|
|
||||||
|
local nodes = util.get_nodes(backend.endpoints)
|
||||||
|
local hash_salt = backend["name"]
|
||||||
|
|
||||||
|
-- override sticky.balancer methods
|
||||||
|
o.instance = util_nodemap:new(nodes, hash_salt)
|
||||||
|
o.get_routing_key = get_routing_key
|
||||||
|
o.set_routing_key = set_routing_key
|
||||||
|
o.pick_new_upstream = pick_new_upstream
|
||||||
|
|
||||||
|
return sticky_balancer
|
||||||
|
end
|
||||||
|
|
||||||
|
return _M
|
|
@ -1,8 +1,8 @@
|
||||||
|
local affinity_balanced = require("affinity.balanced")
|
||||||
|
local affinity_persistent = require("affinity.persistent")
|
||||||
local balancer_resty = require("balancer.resty")
|
local balancer_resty = require("balancer.resty")
|
||||||
local resty_chash = require("resty.chash")
|
|
||||||
local util = require("util")
|
local util = require("util")
|
||||||
local ck = require("resty.cookie")
|
local ck = require("resty.cookie")
|
||||||
local math = require("math")
|
|
||||||
local ngx_balancer = require("ngx.balancer")
|
local ngx_balancer = require("ngx.balancer")
|
||||||
local split = require("util.split")
|
local split = require("util.split")
|
||||||
|
|
||||||
|
@ -10,34 +10,60 @@ local string_format = string.format
|
||||||
local ngx_log = ngx.log
|
local ngx_log = ngx.log
|
||||||
local INFO = ngx.INFO
|
local INFO = ngx.INFO
|
||||||
|
|
||||||
local _M = balancer_resty:new({ factory = resty_chash, name = "sticky" })
|
local _M = balancer_resty:new({ name = "sticky" })
|
||||||
local DEFAULT_COOKIE_NAME = "route"
|
local DEFAULT_COOKIE_NAME = "route"
|
||||||
|
|
||||||
-- Consider the situation of N upstreams one of which is failing.
|
|
||||||
-- Then the probability to obtain failing upstream after M iterations would be close to (1/N)**M.
|
|
||||||
-- For the worst case (2 upstreams; 20 iterations) it would be ~10**(-6)
|
|
||||||
-- which is much better then ~10**(-3) for 10 iterations.
|
|
||||||
local MAX_UPSTREAM_CHECKS_COUNT = 20
|
|
||||||
|
|
||||||
function _M.cookie_name(self)
|
function _M.cookie_name(self)
|
||||||
return self.cookie_session_affinity.name or DEFAULT_COOKIE_NAME
|
return self.cookie_session_affinity.name or DEFAULT_COOKIE_NAME
|
||||||
end
|
end
|
||||||
|
|
||||||
function _M.new(self, backend)
|
local function init_affinity_mode(self, backend)
|
||||||
local nodes = util.get_nodes(backend.endpoints)
|
local mode = backend["sessionAffinityConfig"]["mode"] or 'balanced'
|
||||||
|
|
||||||
|
-- set default mode to 'balanced' for backwards compatibility
|
||||||
|
if mode == nil or mode == '' then
|
||||||
|
mode = 'balanced'
|
||||||
|
end
|
||||||
|
|
||||||
|
self.affinity_mode = mode
|
||||||
|
|
||||||
|
if mode == 'persistent' then
|
||||||
|
return affinity_persistent:new(self, backend)
|
||||||
|
end
|
||||||
|
|
||||||
|
-- default is 'balanced' for backwards compatibility
|
||||||
|
if mode ~= 'balanced' then
|
||||||
|
ngx.log(ngx.WARN, string.format("Invalid affinity mode '%s'! Using 'balanced' as a default.", mode))
|
||||||
|
end
|
||||||
|
|
||||||
|
return affinity_balanced:new(self, backend)
|
||||||
|
end
|
||||||
|
|
||||||
|
function _M.new(self, backend)
|
||||||
local o = {
|
local o = {
|
||||||
instance = self.factory:new(nodes),
|
instance = nil,
|
||||||
|
affinity_mode = nil,
|
||||||
traffic_shaping_policy = backend.trafficShapingPolicy,
|
traffic_shaping_policy = backend.trafficShapingPolicy,
|
||||||
alternative_backends = backend.alternativeBackends,
|
alternative_backends = backend.alternativeBackends,
|
||||||
cookie_session_affinity = backend["sessionAffinityConfig"]["cookieSessionAffinity"]
|
cookie_session_affinity = backend["sessionAffinityConfig"]["cookieSessionAffinity"]
|
||||||
}
|
}
|
||||||
setmetatable(o, self)
|
setmetatable(o, self)
|
||||||
self.__index = self
|
self.__index = self
|
||||||
return o
|
|
||||||
|
return init_affinity_mode(o, backend)
|
||||||
end
|
end
|
||||||
|
|
||||||
local function set_cookie(self, value)
|
function _M.get_cookie(self)
|
||||||
|
local cookie, err = ck:new()
|
||||||
|
if not cookie then
|
||||||
|
ngx.log(ngx.ERR, err)
|
||||||
|
end
|
||||||
|
|
||||||
|
return cookie:get(self:cookie_name())
|
||||||
|
end
|
||||||
|
|
||||||
|
function _M.set_cookie(self, value)
|
||||||
local cookie, err = ck:new()
|
local cookie, err = ck:new()
|
||||||
if not cookie then
|
if not cookie then
|
||||||
ngx.log(ngx.ERR, err)
|
ngx.log(ngx.ERR, err)
|
||||||
|
@ -86,19 +112,30 @@ local function get_failed_upstreams()
|
||||||
return indexed_upstream_addrs
|
return indexed_upstream_addrs
|
||||||
end
|
end
|
||||||
|
|
||||||
local function pick_new_upstream(self)
|
--- get_routing_key gets the current routing key from the cookie
|
||||||
local failed_upstreams = get_failed_upstreams()
|
-- @treturn string, string The routing key and an error message if an error occured.
|
||||||
|
function _M.get_routing_key(self)
|
||||||
|
-- interface method to get the routing key from the cookie
|
||||||
|
-- has to be overridden by an affinity mode
|
||||||
|
ngx.log(ngx.ERR, "[BUG] Failed to get routing key as no implementation has been provided!")
|
||||||
|
return nil, nil
|
||||||
|
end
|
||||||
|
|
||||||
for i = 1, MAX_UPSTREAM_CHECKS_COUNT do
|
--- set_routing_key sets the current routing key on the cookie
|
||||||
local key = string.format("%s.%s.%s", ngx.now() + i, ngx.worker.pid(), math.random(999999))
|
-- @tparam string key The routing key to set on the cookie.
|
||||||
|
function _M.set_routing_key(self, key)
|
||||||
local new_upstream = self.instance:find(key)
|
-- interface method to set the routing key on the cookie
|
||||||
|
-- has to be overridden by an affinity mode
|
||||||
if not failed_upstreams[new_upstream] then
|
ngx.log(ngx.ERR, "[BUG] Failed to set routing key as no implementation has been provided!")
|
||||||
return new_upstream, key
|
end
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
|
--- pick_new_upstream picks a new upstream while ignoring the given failed upstreams.
|
||||||
|
-- @tparam {[string]=boolean} A table of upstreams to ignore where the key is the endpoint and the value a boolean.
|
||||||
|
-- @treturn string, string The endpoint and its key.
|
||||||
|
function _M.pick_new_upstream(self, failed_upstreams)
|
||||||
|
-- interface method to get a new upstream
|
||||||
|
-- has to be overridden by an affinity mode
|
||||||
|
ngx.log(ngx.ERR, "[BUG] Failed to pick new upstream as no implementation has been provided!")
|
||||||
return nil, nil
|
return nil, nil
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -128,15 +165,9 @@ local function should_set_cookie(self)
|
||||||
end
|
end
|
||||||
|
|
||||||
function _M.balance(self)
|
function _M.balance(self)
|
||||||
local cookie, err = ck:new()
|
|
||||||
if not cookie then
|
|
||||||
ngx.log(ngx.ERR, "error while initializing cookie: " .. tostring(err))
|
|
||||||
return
|
|
||||||
end
|
|
||||||
|
|
||||||
local upstream_from_cookie
|
local upstream_from_cookie
|
||||||
|
|
||||||
local key = cookie:get(self:cookie_name())
|
local key = self:get_routing_key()
|
||||||
if key then
|
if key then
|
||||||
upstream_from_cookie = self.instance:find(key)
|
upstream_from_cookie = self.instance:find(key)
|
||||||
end
|
end
|
||||||
|
@ -151,24 +182,34 @@ function _M.balance(self)
|
||||||
|
|
||||||
local new_upstream
|
local new_upstream
|
||||||
|
|
||||||
new_upstream, key = pick_new_upstream(self)
|
new_upstream, key = self:pick_new_upstream(get_failed_upstreams())
|
||||||
if not new_upstream then
|
if not new_upstream then
|
||||||
ngx.log(ngx.WARN, string.format("failed to get new upstream; using upstream %s", new_upstream))
|
ngx.log(ngx.WARN, string.format("failed to get new upstream; using upstream %s", new_upstream))
|
||||||
elseif should_set_cookie(self) then
|
elseif should_set_cookie(self) then
|
||||||
set_cookie(self, key)
|
self:set_routing_key(key)
|
||||||
end
|
end
|
||||||
|
|
||||||
return new_upstream
|
return new_upstream
|
||||||
end
|
end
|
||||||
|
|
||||||
function _M.sync(self, backend)
|
function _M.sync(self, backend)
|
||||||
|
local changed = false
|
||||||
|
|
||||||
|
-- check and reinit affinity mode before syncing the balancer which will reinit the nodes
|
||||||
|
if self.affinity_mode ~= backend.sessionAffinityConfig.mode then
|
||||||
|
changed = true
|
||||||
|
init_affinity_mode(self, backend)
|
||||||
|
end
|
||||||
|
|
||||||
|
-- reload balancer nodes
|
||||||
balancer_resty.sync(self, backend)
|
balancer_resty.sync(self, backend)
|
||||||
|
|
||||||
-- Reload the balancer if any of the annotations have changed.
|
-- Reload the balancer if any of the annotations have changed.
|
||||||
local changed = not util.deep_compare(
|
changed = changed or not util.deep_compare(
|
||||||
self.cookie_session_affinity,
|
self.cookie_session_affinity,
|
||||||
backend.sessionAffinityConfig.cookieSessionAffinity
|
backend.sessionAffinityConfig.cookieSessionAffinity
|
||||||
)
|
)
|
||||||
|
|
||||||
if not changed then
|
if not changed then
|
||||||
return
|
return
|
||||||
end
|
end
|
||||||
|
|
|
@ -15,11 +15,16 @@ local function reset_ngx()
|
||||||
end
|
end
|
||||||
|
|
||||||
function get_mocked_cookie_new()
|
function get_mocked_cookie_new()
|
||||||
return function(self)
|
local o = { value = nil }
|
||||||
return {
|
local mock = {
|
||||||
get = function(self, n) return nil, "error" end,
|
get = function(self, n) return self.value end,
|
||||||
set = function(self, n) return true, "" end
|
set = function(self, c) self.value = c.value ; return true, nil end
|
||||||
}
|
}
|
||||||
|
setmetatable(o, mock)
|
||||||
|
mock.__index = mock
|
||||||
|
|
||||||
|
return function(self)
|
||||||
|
return o;
|
||||||
end
|
end
|
||||||
end
|
end
|
||||||
|
|
||||||
|
@ -229,7 +234,7 @@ describe("Sticky", function()
|
||||||
end)
|
end)
|
||||||
end)
|
end)
|
||||||
|
|
||||||
local function get_several_test_backends(change_on_failure)
|
local function get_several_test_backends(option)
|
||||||
return {
|
return {
|
||||||
name = "access-router-production-web-80",
|
name = "access-router-production-web-80",
|
||||||
endpoints = {
|
endpoints = {
|
||||||
|
@ -238,7 +243,13 @@ describe("Sticky", function()
|
||||||
},
|
},
|
||||||
sessionAffinityConfig = {
|
sessionAffinityConfig = {
|
||||||
name = "cookie",
|
name = "cookie",
|
||||||
cookieSessionAffinity = { name = "test_name", hash = "sha1", change_on_failure = change_on_failure }
|
mode = option["mode"],
|
||||||
|
cookieSessionAffinity = {
|
||||||
|
name = "test_name",
|
||||||
|
hash = "sha1",
|
||||||
|
change_on_failure = option["change_on_failure"],
|
||||||
|
locations = { ['test.com'] = {'/'} }
|
||||||
|
}
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
end
|
end
|
||||||
|
@ -257,21 +268,20 @@ describe("Sticky", function()
|
||||||
|
|
||||||
context("when request to upstream fails", function()
|
context("when request to upstream fails", function()
|
||||||
it("changes upstream when change_on_failure option is true", function()
|
it("changes upstream when change_on_failure option is true", function()
|
||||||
-- create sticky cookie
|
local options = {
|
||||||
cookie.new = function(self)
|
{["change_on_failure"] = false, ["mode"] = nil},
|
||||||
local return_obj = {
|
{["change_on_failure"] = false, ["mode"] = 'balanced'},
|
||||||
set = function(v) return false, nil end,
|
{["change_on_failure"] = false, ["mode"] = 'persistent'},
|
||||||
get = function(k) return "" end,
|
{["change_on_failure"] = true, ["mode"] = nil},
|
||||||
|
{["change_on_failure"] = true, ["mode"] = 'balanced'},
|
||||||
|
{["change_on_failure"] = true, ["mode"] = 'persistent'}
|
||||||
}
|
}
|
||||||
return return_obj, false
|
|
||||||
end
|
|
||||||
|
|
||||||
local options = {false, true}
|
|
||||||
|
|
||||||
for _, option in ipairs(options) do
|
for _, option in ipairs(options) do
|
||||||
local sticky_balancer_instance = sticky:new(get_several_test_backends(option))
|
local sticky_balancer_instance = sticky:new(get_several_test_backends(option))
|
||||||
|
|
||||||
local old_upstream = sticky_balancer_instance:balance()
|
local old_upstream = sticky_balancer_instance:balance()
|
||||||
|
assert.is.Not.Nil(old_upstream)
|
||||||
for _ = 1, 100 do
|
for _ = 1, 100 do
|
||||||
-- make sure upstream doesn't change on subsequent calls of balance()
|
-- make sure upstream doesn't change on subsequent calls of balance()
|
||||||
assert.equal(old_upstream, sticky_balancer_instance:balance())
|
assert.equal(old_upstream, sticky_balancer_instance:balance())
|
||||||
|
@ -281,11 +291,11 @@ describe("Sticky", function()
|
||||||
sticky_balancer_instance.get_last_failure = function()
|
sticky_balancer_instance.get_last_failure = function()
|
||||||
return "failed"
|
return "failed"
|
||||||
end
|
end
|
||||||
_G.ngx.var = { upstream_addr = old_upstream }
|
_G.ngx.var.upstream_addr = old_upstream
|
||||||
|
|
||||||
for _ = 1, 100 do
|
for _ = 1, 100 do
|
||||||
local new_upstream = sticky_balancer_instance:balance()
|
local new_upstream = sticky_balancer_instance:balance()
|
||||||
if option == false then
|
if option["change_on_failure"] == false then
|
||||||
-- upstream should be the same inspite of error, if change_on_failure option is false
|
-- upstream should be the same inspite of error, if change_on_failure option is false
|
||||||
assert.equal(new_upstream, old_upstream)
|
assert.equal(new_upstream, old_upstream)
|
||||||
else
|
else
|
||||||
|
|
167
rootfs/etc/nginx/lua/test/util/nodemap_test.lua
Normal file
167
rootfs/etc/nginx/lua/test/util/nodemap_test.lua
Normal file
|
@ -0,0 +1,167 @@
|
||||||
|
local util = require("util")
|
||||||
|
local nodemap = require("util.nodemap")
|
||||||
|
|
||||||
|
local function get_test_backend_single()
|
||||||
|
return {
|
||||||
|
name = "access-router-production-web-80",
|
||||||
|
endpoints = {
|
||||||
|
{ address = "10.184.7.40", port = "8080", maxFails = 0, failTimeout = 0 }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
end
|
||||||
|
|
||||||
|
local function get_test_backend_multi()
|
||||||
|
return {
|
||||||
|
name = "access-router-production-web-80",
|
||||||
|
endpoints = {
|
||||||
|
{ address = "10.184.7.40", port = "8080", maxFails = 0, failTimeout = 0 },
|
||||||
|
{ address = "10.184.7.41", port = "8080", maxFails = 0, failTimeout = 0 }
|
||||||
|
}
|
||||||
|
}
|
||||||
|
end
|
||||||
|
|
||||||
|
local function get_test_nodes_ignore(endpoint)
|
||||||
|
local ignore = {}
|
||||||
|
ignore[endpoint] = true
|
||||||
|
return ignore
|
||||||
|
end
|
||||||
|
|
||||||
|
describe("Node Map", function()
|
||||||
|
|
||||||
|
local test_backend_single = get_test_backend_single()
|
||||||
|
local test_backend_multi = get_test_backend_multi()
|
||||||
|
local test_salt = test_backend_single.name
|
||||||
|
local test_nodes_single = util.get_nodes(test_backend_single.endpoints)
|
||||||
|
local test_nodes_multi = util.get_nodes(test_backend_multi.endpoints)
|
||||||
|
local test_endpoint1 = test_backend_multi.endpoints[1].address .. ":" .. test_backend_multi.endpoints[1].port
|
||||||
|
local test_endpoint2 = test_backend_multi.endpoints[2].address .. ":" .. test_backend_multi.endpoints[2].port
|
||||||
|
local test_nodes_ignore = get_test_nodes_ignore(test_endpoint1)
|
||||||
|
|
||||||
|
describe("new()", function()
|
||||||
|
context("when no salt has been provided", function()
|
||||||
|
it("random() returns an unsalted key", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_single, nil)
|
||||||
|
local expected_endpoint = test_endpoint1
|
||||||
|
local expected_hash_key = ngx.md5(expected_endpoint)
|
||||||
|
local actual_endpoint
|
||||||
|
local actual_hash_key
|
||||||
|
|
||||||
|
actual_endpoint, actual_hash_key = nodemap_instance:random()
|
||||||
|
|
||||||
|
assert.equal(actual_endpoint, expected_endpoint)
|
||||||
|
assert.equal(expected_hash_key, actual_hash_key)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when a salt has been provided", function()
|
||||||
|
it("random() returns a salted key", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||||
|
local expected_endpoint = test_endpoint1
|
||||||
|
local expected_hash_key = ngx.md5(test_salt .. expected_endpoint)
|
||||||
|
local actual_endpoint
|
||||||
|
local actual_hash_key
|
||||||
|
|
||||||
|
actual_endpoint, actual_hash_key = nodemap_instance:random()
|
||||||
|
|
||||||
|
assert.equal(actual_endpoint, expected_endpoint)
|
||||||
|
assert.equal(expected_hash_key, actual_hash_key)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when no nodes have been provided", function()
|
||||||
|
it("random() returns nil", function()
|
||||||
|
local nodemap_instance = nodemap:new({}, test_salt)
|
||||||
|
local actual_endpoint
|
||||||
|
local actual_hash_key
|
||||||
|
|
||||||
|
actual_endpoint, actual_hash_key = nodemap_instance:random()
|
||||||
|
|
||||||
|
assert.equal(actual_endpoint, nil)
|
||||||
|
assert.equal(expected_hash_key, nil)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
describe("find()", function()
|
||||||
|
before_each(function()
|
||||||
|
package.loaded["util.nodemap"] = nil
|
||||||
|
nodemap = require("util.nodemap")
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when a hash key is valid", function()
|
||||||
|
it("find() returns the correct endpoint", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||||
|
local test_hash_key
|
||||||
|
local expected_endpoint
|
||||||
|
local actual_endpoint
|
||||||
|
|
||||||
|
expected_endpoint, test_hash_key = nodemap_instance:random()
|
||||||
|
assert.not_equal(expected_endpoint, nil)
|
||||||
|
assert.not_equal(test_hash_key, nil)
|
||||||
|
|
||||||
|
actual_endpoint = nodemap_instance:find(test_hash_key)
|
||||||
|
assert.equal(actual_endpoint, expected_endpoint)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when a hash key is invalid", function()
|
||||||
|
it("find() returns nil", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||||
|
local test_hash_key = "invalid or nonexistent hash key"
|
||||||
|
local actual_endpoint
|
||||||
|
|
||||||
|
actual_endpoint = nodemap_instance:find(test_hash_key)
|
||||||
|
|
||||||
|
assert.equal(actual_endpoint, nil)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
|
||||||
|
describe("random_except()", function()
|
||||||
|
before_each(function()
|
||||||
|
package.loaded["util.nodemap"] = nil
|
||||||
|
nodemap = require("util.nodemap")
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when nothing has been excluded", function()
|
||||||
|
it("random_except() returns the correct endpoint", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||||
|
local expected_endpoint = test_endpoint1
|
||||||
|
local test_hash_key
|
||||||
|
local actual_endpoint
|
||||||
|
|
||||||
|
actual_endpoint, test_hash_key = nodemap_instance:random_except({})
|
||||||
|
assert.equal(expected_endpoint, actual_endpoint)
|
||||||
|
assert.not_equal(test_hash_key, nil)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when everything has been excluded", function()
|
||||||
|
it("random_except() returns nil", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_single, test_salt)
|
||||||
|
local actual_hash_key
|
||||||
|
local actual_endpoint
|
||||||
|
|
||||||
|
actual_endpoint, actual_hash_key = nodemap_instance:random_except(test_nodes_ignore)
|
||||||
|
|
||||||
|
assert.equal(actual_endpoint, nil)
|
||||||
|
assert.equal(actual_hash_key, nil)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
|
||||||
|
context("when an endpoint has been excluded", function()
|
||||||
|
it("random_except() does not return it", function()
|
||||||
|
local nodemap_instance = nodemap:new(test_nodes_multi, test_salt)
|
||||||
|
local expected_endpoint = test_endpoint2
|
||||||
|
local actual_endpoint
|
||||||
|
local test_hash_key
|
||||||
|
|
||||||
|
actual_endpoint, test_hash_key = nodemap_instance:random_except(test_nodes_ignore)
|
||||||
|
|
||||||
|
assert.equal(actual_endpoint, expected_endpoint)
|
||||||
|
assert.not_equal(test_hash_key, nil)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
end)
|
||||||
|
end)
|
120
rootfs/etc/nginx/lua/util/nodemap.lua
Normal file
120
rootfs/etc/nginx/lua/util/nodemap.lua
Normal file
|
@ -0,0 +1,120 @@
|
||||||
|
local math = require("math")
|
||||||
|
local util = require("util")
|
||||||
|
|
||||||
|
local _M = {}
|
||||||
|
|
||||||
|
--- create_map generates the node hash table
|
||||||
|
-- @tparam {[string]=number} nodes A table with the node as a key and its weight as a value.
|
||||||
|
-- @tparam string salt A salt that will be used to generate salted hash keys.
|
||||||
|
local function create_map(nodes, salt)
|
||||||
|
local hash_map = {}
|
||||||
|
|
||||||
|
for endpoint, _ in pairs(nodes) do
|
||||||
|
-- obfuscate the endpoint with a shared key to prevent brute force
|
||||||
|
-- and rainbow table attacks which could reveal internal endpoints
|
||||||
|
local key = salt .. endpoint
|
||||||
|
local hash_key = ngx.md5(key)
|
||||||
|
hash_map[hash_key] = endpoint
|
||||||
|
end
|
||||||
|
|
||||||
|
return hash_map
|
||||||
|
end
|
||||||
|
|
||||||
|
--- get_random_node picks a random node from the given map.
|
||||||
|
-- @tparam {[string], ...} map A key to node hash table.
|
||||||
|
-- @treturn string,string The node and its key
|
||||||
|
local function get_random_node(map)
|
||||||
|
local size = util.tablelength(map)
|
||||||
|
|
||||||
|
if size < 1 then
|
||||||
|
return nil, nil
|
||||||
|
end
|
||||||
|
|
||||||
|
local index = math.random(1, size)
|
||||||
|
local count = 1
|
||||||
|
|
||||||
|
for key, endpoint in pairs(map) do
|
||||||
|
if count == index then
|
||||||
|
return endpoint, key
|
||||||
|
end
|
||||||
|
|
||||||
|
count = count + 1
|
||||||
|
end
|
||||||
|
|
||||||
|
ngx.log(ngx.ERR, string.format("Failed to find node %d of %d! This is a bug, please report!", index, size))
|
||||||
|
|
||||||
|
return nil, nil
|
||||||
|
end
|
||||||
|
|
||||||
|
--- new constructs a new instance of the node map
|
||||||
|
--
|
||||||
|
-- The map uses MD5 to create hash keys for a given node. For security reasons it supports
|
||||||
|
-- salted hash keys, to prevent attackers from using rainbow tables or brute forcing
|
||||||
|
-- the node endpoints, which would reveal cluster internal network information.
|
||||||
|
--
|
||||||
|
-- To make sure hash keys are reproducible on different ingress controller instances the salt
|
||||||
|
-- needs to be shared and therefore is not simply generated randomly.
|
||||||
|
--
|
||||||
|
-- @tparam {[string]=number} endpoints A table with the node endpoint as a key and its weight as a value.
|
||||||
|
-- @tparam[opt] string hash_salt A optional hash salt that will be used to obfuscate the hash key.
|
||||||
|
function _M.new(self, endpoints, hash_salt)
|
||||||
|
|
||||||
|
if hash_salt == nil then
|
||||||
|
hash_salt = ''
|
||||||
|
end
|
||||||
|
|
||||||
|
-- the endpoints have to be saved as 'nodes' to keep compatibility to balancer.resty
|
||||||
|
local o = {
|
||||||
|
salt = hash_salt,
|
||||||
|
nodes = endpoints,
|
||||||
|
map = create_map(endpoints, hash_salt)
|
||||||
|
}
|
||||||
|
|
||||||
|
setmetatable(o, self)
|
||||||
|
self.__index = self
|
||||||
|
return o
|
||||||
|
end
|
||||||
|
|
||||||
|
--- reinit reinitializes the node map reusing the original salt
|
||||||
|
-- @tparam {[string]=number} nodes A table with the node as a key and its weight as a value.
|
||||||
|
function _M.reinit(self, nodes)
|
||||||
|
self.nodes = nodes
|
||||||
|
self.map = create_map(nodes, self.salt)
|
||||||
|
end
|
||||||
|
|
||||||
|
--- find looks up a node by hash key.
|
||||||
|
-- @tparam string key The hash key.
|
||||||
|
-- @treturn string The node.
|
||||||
|
function _M.find(self, key)
|
||||||
|
return self.map[key]
|
||||||
|
end
|
||||||
|
|
||||||
|
--- random picks a random node from the hashmap.
|
||||||
|
-- @treturn string,string A random node and its key or both nil.
|
||||||
|
function _M.random(self)
|
||||||
|
return get_random_node(self.map)
|
||||||
|
end
|
||||||
|
|
||||||
|
--- random_except picks a random node from the hashmap, ignoring the nodes in the given table
|
||||||
|
-- @tparam {string, } ignore_nodes A table of nodes to ignore, the node needs to be the key,
|
||||||
|
-- the value needs to be set to true
|
||||||
|
-- @treturn string,string A random node and its key or both nil.
|
||||||
|
function _M.random_except(self, ignore_nodes)
|
||||||
|
local valid_nodes = {}
|
||||||
|
|
||||||
|
-- avoid generating the map if no ignores where provided
|
||||||
|
if ignore_nodes == nil or util.tablelength(ignore_nodes) == 0 then
|
||||||
|
return get_random_node(self.map)
|
||||||
|
end
|
||||||
|
|
||||||
|
-- generate valid endpoints
|
||||||
|
for key, endpoint in pairs(self.map) do
|
||||||
|
if not ignore_nodes[endpoint] then
|
||||||
|
valid_nodes[key] = endpoint
|
||||||
|
end
|
||||||
|
end
|
||||||
|
|
||||||
|
return get_random_node(valid_nodes)
|
||||||
|
end
|
||||||
|
|
||||||
|
return _M
|
|
@ -54,6 +54,7 @@
|
||||||
}],
|
}],
|
||||||
"sessionAffinityConfig": {
|
"sessionAffinityConfig": {
|
||||||
"name": "",
|
"name": "",
|
||||||
|
"mode": "",
|
||||||
"cookieSessionAffinity": {
|
"cookieSessionAffinity": {
|
||||||
"name": ""
|
"name": ""
|
||||||
}
|
}
|
||||||
|
@ -126,6 +127,7 @@
|
||||||
}],
|
}],
|
||||||
"sessionAffinityConfig": {
|
"sessionAffinityConfig": {
|
||||||
"name": "",
|
"name": "",
|
||||||
|
"mode": "",
|
||||||
"cookieSessionAffinity": {
|
"cookieSessionAffinity": {
|
||||||
"name": ""
|
"name": ""
|
||||||
}
|
}
|
||||||
|
@ -191,6 +193,7 @@
|
||||||
}],
|
}],
|
||||||
"sessionAffinityConfig": {
|
"sessionAffinityConfig": {
|
||||||
"name": "",
|
"name": "",
|
||||||
|
"mode": "",
|
||||||
"cookieSessionAffinity": {
|
"cookieSessionAffinity": {
|
||||||
"name": ""
|
"name": ""
|
||||||
}
|
}
|
||||||
|
|
|
@ -54,6 +54,7 @@
|
||||||
}],
|
}],
|
||||||
"sessionAffinityConfig": {
|
"sessionAffinityConfig": {
|
||||||
"name": "",
|
"name": "",
|
||||||
|
"mode": "",
|
||||||
"cookieSessionAffinity": {
|
"cookieSessionAffinity": {
|
||||||
"name": ""
|
"name": ""
|
||||||
}
|
}
|
||||||
|
@ -126,6 +127,7 @@
|
||||||
}],
|
}],
|
||||||
"sessionAffinityConfig": {
|
"sessionAffinityConfig": {
|
||||||
"name": "",
|
"name": "",
|
||||||
|
"mode": "",
|
||||||
"cookieSessionAffinity": {
|
"cookieSessionAffinity": {
|
||||||
"name": ""
|
"name": ""
|
||||||
}
|
}
|
||||||
|
@ -191,6 +193,7 @@
|
||||||
}],
|
}],
|
||||||
"sessionAffinityConfig": {
|
"sessionAffinityConfig": {
|
||||||
"name": "",
|
"name": "",
|
||||||
|
"mode": "",
|
||||||
"cookieSessionAffinity": {
|
"cookieSessionAffinity": {
|
||||||
"name": ""
|
"name": ""
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in a new issue