*/
const f64 multipath_next_hop_error_tolerance = 0.1;
-#undef LB_DEBUG
+static const char *load_balance_attr_names[] = LOAD_BALANCE_ATTR_NAMES;
+#undef LB_DEBUG
#ifdef LB_DEBUG
#define LB_DBG(_lb, _fmt, _args...) \
{ \
s = format(s, "[proto:%U ", format_dpo_proto, lb->lb_proto);
s = format(s, "index:%d buckets:%d ", lbi, lb->lb_n_buckets);
s = format(s, "uRPF:%d ", lb->lb_urpf);
+ if (lb->lb_flags)
+ {
+ load_balance_attr_t attr;
+
+ s = format(s, "flags:[");
+
+ FOR_EACH_LOAD_BALANCE_ATTR(attr)
+ {
+ if (lb->lb_flags & (1 << attr))
+ {
+ s = format (s, "%s", load_balance_attr_names[attr]);
+ }
+ }
+ s = format(s, "] ");
+ }
s = format(s, "to:[%Ld:%Ld]", to.packets, to.bytes);
if (0 != via.packets)
{
* next hop adjacencies.
*/
static void
-load_balance_fill_buckets (load_balance_t *lb,
- load_balance_path_t *nhs,
- dpo_id_t *buckets,
- u32 n_buckets)
+load_balance_fill_buckets_norm (load_balance_t *lb,
+ load_balance_path_t *nhs,
+ dpo_id_t *buckets,
+ u32 n_buckets)
{
- load_balance_path_t * nh;
+ load_balance_path_t *nh;
u16 ii, bucket;
bucket = 0;
}
}
}
+static void
+load_balance_fill_buckets_sticky (load_balance_t *lb,
+ load_balance_path_t *nhs,
+ dpo_id_t *buckets,
+ u32 n_buckets)
+{
+ load_balance_path_t *nh, *fwding_paths;
+ u16 ii, bucket, fpath;
+
+ fpath = bucket = 0;
+ fwding_paths = NULL;
+
+ vec_foreach (nh, nhs)
+ {
+ if (!dpo_is_drop(&nh->path_dpo))
+ {
+ vec_add1(fwding_paths, *nh);
+ }
+ }
+ if (vec_len(fwding_paths) == 0)
+ fwding_paths = vec_dup(nhs);
+
+ /*
+ * the next-hops have normalised weights. that means their sum is the number
+ * of buckets we need to fill.
+ */
+ vec_foreach (nh, nhs)
+ {
+ for (ii = 0; ii < nh->path_weight; ii++)
+ {
+ ASSERT(bucket < n_buckets);
+ if (!dpo_is_drop(&nh->path_dpo))
+ {
+ load_balance_set_bucket_i(lb, bucket++, buckets, &nh->path_dpo);
+ }
+ else
+ {
+ /* fill the bucks from the next up path */
+ load_balance_set_bucket_i(lb, bucket++, buckets, &fwding_paths[fpath].path_dpo);
+ fpath = (fpath + 1) % vec_len(fwding_paths);
+ }
+ }
+ }
+
+ vec_free(fwding_paths);
+}
+
+static void
+load_balance_fill_buckets (load_balance_t *lb,
+ load_balance_path_t *nhs,
+ dpo_id_t *buckets,
+ u32 n_buckets,
+ load_balance_flags_t flags)
+{
+ if (flags & LOAD_BALANCE_FLAG_STICKY)
+ {
+ load_balance_fill_buckets_sticky(lb, nhs, buckets, n_buckets);
+ }
+ else
+ {
+ load_balance_fill_buckets_norm(lb, nhs, buckets, n_buckets);
+ }
+}
static inline void
load_balance_set_n_buckets (load_balance_t *lb,
ASSERT(DPO_LOAD_BALANCE == dpo->dpoi_type);
lb = load_balance_get(dpo->dpoi_index);
+ lb->lb_flags = flags;
fixed_nhs = load_balance_multipath_next_hop_fixup(raw_nhs, lb->lb_proto);
n_buckets =
ip_multipath_normalize_next_hops((NULL == fixed_nhs ?
load_balance_fill_buckets(lb, nhs,
load_balance_get_buckets(lb),
- n_buckets);
+ n_buckets, flags);
lb->lb_map = lbmi;
}
else
*/
load_balance_fill_buckets(lb, nhs,
load_balance_get_buckets(lb),
- n_buckets);
+ n_buckets, flags);
lb->lb_map = lbmi;
}
else if (n_buckets > lb->lb_n_buckets)
load_balance_fill_buckets(lb, nhs,
lb->lb_buckets,
- n_buckets);
+ n_buckets, flags);
CLIB_MEMORY_BARRIER();
load_balance_set_n_buckets(lb, n_buckets);
*/
load_balance_fill_buckets(lb, nhs,
load_balance_get_buckets(lb),
- n_buckets);
+ n_buckets, flags);
CLIB_MEMORY_BARRIER();
load_balance_set_n_buckets(lb, n_buckets);
}
n_buckets - 1,
CLIB_CACHE_LINE_BYTES);
- load_balance_fill_buckets(lb, nhs, new_buckets, n_buckets);
+ load_balance_fill_buckets(lb, nhs, new_buckets,
+ n_buckets, flags);
CLIB_MEMORY_BARRIER();
lb->lb_buckets = new_buckets;
CLIB_MEMORY_BARRIER();
*/
load_balance_fill_buckets(lb, nhs,
lb->lb_buckets_inline,
- n_buckets);
+ n_buckets, flags);
CLIB_MEMORY_BARRIER();
load_balance_set_n_buckets(lb, n_buckets);
CLIB_MEMORY_BARRIER();
load_balance_set_n_buckets(lb, n_buckets);
CLIB_MEMORY_BARRIER();
- load_balance_fill_buckets(lb, nhs,
- buckets,
- n_buckets);
+ load_balance_fill_buckets(lb, nhs, buckets,
+ n_buckets, flags);
for (ii = n_buckets; ii < old_n_buckets; ii++)
{
if (INDEX_INVALID != lbi)
{
- vlib_cli_output (vm, "%U", format_load_balance, lbi,
+ if (pool_is_free_index(load_balance_pool, lbi))
+ {
+ vlib_cli_output (vm, "no such load-balance:%d", lbi);
+ }
+ else
+ {
+ vlib_cli_output (vm, "%U", format_load_balance, lbi,
LOAD_BALANCE_FORMAT_DETAIL);
+ }
}
else
{
u32 path_weight;
} load_balance_path_t;
+/**
+ * Flags controlling load-balance creation and modification
+ */
+typedef enum load_balance_attr_t_ {
+ LOAD_BALANCE_ATTR_USES_MAP = 0,
+ LOAD_BALANCE_ATTR_STICKY = 1,
+} load_balance_attr_t;
+
+#define LOAD_BALANCE_ATTR_NAMES { \
+ [LOAD_BALANCE_ATTR_USES_MAP] = "uses-map", \
+ [LOAD_BALANCE_ATTR_STICKY] = "sticky", \
+}
+
+#define FOR_EACH_LOAD_BALANCE_ATTR(_attr) \
+ for (_attr = 0; _attr <= LOAD_BALANCE_ATTR_STICKY; _attr++)
+
+typedef enum load_balance_flags_t_ {
+ LOAD_BALANCE_FLAG_NONE = 0,
+ LOAD_BALANCE_FLAG_USES_MAP = (1 << 0),
+ LOAD_BALANCE_FLAG_STICKY = (1 << 1),
+} __attribute__((packed)) load_balance_flags_t;
+
/**
* The FIB DPO provieds;
* - load-balancing over the next DPOs in the chain/graph
*/
dpo_proto_t lb_proto;
+ /**
+ * Flags concenring the LB's creation and modification
+ */
+ load_balance_flags_t lb_flags;
+
/**
* Flags from the load-balance's associated fib_entry_t
*/
LOAD_BALANCE_FORMAT_DETAIL = (1 << 0),
} load_balance_format_flags_t;
-/**
- * Flags controlling load-balance creation and modification
- */
-typedef enum load_balance_flags_t_ {
- LOAD_BALANCE_FLAG_NONE = 0,
- LOAD_BALANCE_FLAG_USES_MAP = (1 << 0),
-} load_balance_flags_t;
-
extern index_t load_balance_create(u32 num_buckets,
dpo_proto_t lb_proto,
flow_hash_config_t fhc);
ASSERT(path);
+ vec_add2(hash_key, mnh, 1);
+
+ mnh->path_weight = path->fp_weight;
+ mnh->path_index = path_index;
+
if (fib_path_is_resolved(path_index))
{
- vec_add2(hash_key, mnh, 1);
-
- mnh->path_weight = path->fp_weight;
- mnh->path_index = path_index;
- fib_path_contribute_forwarding(path_index, fct, &mnh->path_dpo);
+ fib_path_contribute_forwarding(path_index, fct, &mnh->path_dpo);
+ }
+ else
+ {
+ dpo_copy(&mnh->path_dpo,
+ drop_dpo_get(fib_forw_chain_type_to_dpo_proto(fct)));
}
-
return (hash_key);
}
fib_path_list_destroy(path_list);
}
+static load_balance_flags_t
+fib_path_list_fwd_flags_2_load_balance (fib_path_list_fwd_flags_t pl_flags)
+{
+ load_balance_flags_t lb_flags = LOAD_BALANCE_FLAG_NONE;
+
+ if (pl_flags & FIB_PATH_LIST_FWD_FLAG_STICKY)
+ {
+ lb_flags |= LOAD_BALANCE_ATTR_STICKY;
+ }
+ return (lb_flags);
+}
+
/*
* fib_path_mk_lb
*
static void
fib_path_list_mk_lb (fib_path_list_t *path_list,
fib_forward_chain_type_t fct,
- dpo_id_t *dpo)
+ dpo_id_t *dpo,
+ fib_path_list_fwd_flags_t flags)
{
load_balance_path_t *nhs;
fib_node_index_t *path_index;
*/
vec_foreach (path_index, path_list->fpl_paths)
{
- nhs = fib_path_append_nh_for_multipath_hash(*path_index,
- fct,
- nhs);
+ if ((flags & FIB_PATH_LIST_FWD_FLAG_STICKY) ||
+ fib_path_is_resolved(*path_index))
+ {
+ nhs = fib_path_append_nh_for_multipath_hash(*path_index,
+ fct, nhs);
+ }
}
/*
load_balance_create(vec_len(nhs),
fib_forw_chain_type_to_dpo_proto(fct),
0 /* FIXME FLOW HASH */));
- load_balance_multipath_update(dpo, nhs, LOAD_BALANCE_FLAG_NONE);
+ load_balance_multipath_update(dpo, nhs,
+ fib_path_list_fwd_flags_2_load_balance(flags));
FIB_PATH_LIST_DBG(path_list, "mk lb: %d", dpo->dpoi_index);
path_list = fib_path_list_get(path_list_index);
- fib_path_list_mk_lb(path_list, fct, dpo);
+ fib_path_list_mk_lb(path_list, fct, dpo, flags);
ASSERT(DPO_LOAD_BALANCE == dpo->dpoi_type);
{
FIB_PATH_LIST_FWD_FLAG_NONE = 0,
FIB_PATH_LIST_FWD_FLAG_COLLAPSE = (1 << 0),
+ FIB_PATH_LIST_FWD_FLAG_STICKY = (1 << 1),
} fib_path_list_fwd_flags_t;
extern void fib_path_list_contribute_forwarding(fib_node_index_t path_list_index,
res = 0;
va_start(ap, n_buckets);
- if (FIB_TEST_I((DPO_LOAD_BALANCE == dpo->dpoi_type),
- "Entry links to %U",
- format_dpo_type, dpo->dpoi_type))
+ if (!FIB_TEST_I((DPO_LOAD_BALANCE == dpo->dpoi_type),
+ "Entry links to %U",
+ format_dpo_type, dpo->dpoi_type))
{
lb = load_balance_get(dpo->dpoi_index);
}
else
{
- res = 0;
+ res = 1;
}
va_end(ap);
return (res);
}
+static int
+fib_test_sticky (void)
+{
+ fib_route_path_t *r_paths = NULL;
+ test_main_t *tm = &test_main;
+ u32 ii, lb_count, pl_count;
+ dpo_id_t dpo = DPO_INVALID;
+ fib_node_index_t pl_index;
+ int res = 0;
+#define N_PATHS 16
+
+ fib_test_lb_bucket_t buckets[N_PATHS];
+ bfd_session_t bfds[N_PATHS] = {{0}};
+
+ lb_count = pool_elts(load_balance_pool);
+ pl_count = fib_path_list_pool_size();
+
+ for (ii = 0; ii < N_PATHS; ii++)
+ {
+ ip46_address_t nh = {
+ .ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02 + ii),
+ };
+ adj_index_t ai;
+
+ ai = adj_nbr_add_or_lock(FIB_PROTOCOL_IP4,
+ VNET_LINK_IP4,
+ &nh, tm->hw[0]->sw_if_index);
+
+ buckets[ii].type = FT_LB_ADJ;
+ buckets[ii].adj.adj = ai;
+
+ bfds[ii].udp.key.peer_addr = nh;
+ bfds[ii].udp.key.sw_if_index = tm->hw[0]->sw_if_index;
+ bfds[ii].hop_type = BFD_HOP_TYPE_SINGLE;
+ bfds[ii].local_state = BFD_STATE_init;
+ adj_bfd_notify(BFD_LISTEN_EVENT_CREATE, &bfds[ii]);
+ bfds[ii].local_state = BFD_STATE_up;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[ii]);
+ }
+
+ for (ii = 0; ii < N_PATHS; ii++)
+ {
+ fib_route_path_t r_path = {
+ .frp_proto = DPO_PROTO_IP4,
+ .frp_addr = {
+ .ip4.as_u32 = clib_host_to_net_u32(0x0a0a0a02 + ii),
+ },
+ .frp_sw_if_index = tm->hw[0]->sw_if_index,
+ .frp_weight = 1,
+ .frp_fib_index = ~0,
+ };
+ vec_add1(r_paths, r_path);
+ };
+
+ pl_index = fib_path_list_create(FIB_PATH_LIST_FLAG_SHARED, r_paths);
+ fib_path_list_lock(pl_index);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[0],
+ &buckets[1],
+ &buckets[2],
+ &buckets[3],
+ &buckets[4],
+ &buckets[5],
+ &buckets[6],
+ &buckets[7],
+ &buckets[8],
+ &buckets[9],
+ &buckets[10],
+ &buckets[11],
+ &buckets[12],
+ &buckets[13],
+ &buckets[14],
+ &buckets[15]),
+ "Setup OK");
+
+ /* take down paths */
+ bfds[0].local_state = BFD_STATE_down;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[0]);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[1],
+ &buckets[1],
+ &buckets[2],
+ &buckets[3],
+ &buckets[4],
+ &buckets[5],
+ &buckets[6],
+ &buckets[7],
+ &buckets[8],
+ &buckets[9],
+ &buckets[10],
+ &buckets[11],
+ &buckets[12],
+ &buckets[13],
+ &buckets[14],
+ &buckets[15]),
+ "Failed at shut-down path 0");
+
+ bfds[7].local_state = BFD_STATE_down;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[7]);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[1],
+ &buckets[1],
+ &buckets[2],
+ &buckets[3],
+ &buckets[4],
+ &buckets[5],
+ &buckets[6],
+ &buckets[2],
+ &buckets[8],
+ &buckets[9],
+ &buckets[10],
+ &buckets[11],
+ &buckets[12],
+ &buckets[13],
+ &buckets[14],
+ &buckets[15]),
+ "Failed at shut-down path 7");
+
+ /* paths back up */
+ bfds[0].local_state = BFD_STATE_up;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[0]);
+ bfds[7].local_state = BFD_STATE_up;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[7]);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[0],
+ &buckets[1],
+ &buckets[2],
+ &buckets[3],
+ &buckets[4],
+ &buckets[5],
+ &buckets[6],
+ &buckets[7],
+ &buckets[8],
+ &buckets[9],
+ &buckets[10],
+ &buckets[11],
+ &buckets[12],
+ &buckets[13],
+ &buckets[14],
+ &buckets[15]),
+ "recovery OK");
+
+ fib_path_list_unlock(pl_index);
+
+ /*
+ * non-power of 2 number of buckets
+ */
+ fib_route_path_t *r_paths2 = NULL;
+
+ r_paths2 = vec_dup(r_paths);
+ _vec_len(r_paths2) = 3;
+
+ pl_index = fib_path_list_create(FIB_PATH_LIST_FLAG_SHARED, r_paths2);
+ fib_path_list_lock(pl_index);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[1],
+ &buckets[1],
+ &buckets[1],
+ &buckets[1],
+ &buckets[1],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2]),
+ "non-power of 2");
+
+ bfds[1].local_state = BFD_STATE_down;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[1]);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ /*
+ * path 1's buckets alternate between path 0 and 2
+ */
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[2],
+ &buckets[0],
+ &buckets[2],
+ &buckets[0],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2]),
+ "non-power of 2");
+ bfds[1].local_state = BFD_STATE_up;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[1]);
+
+ fib_path_list_unlock(pl_index);
+
+ /*
+ * unequal cost
+ */
+ fib_route_path_t *r_paths3 = NULL;
+
+ r_paths3 = vec_dup(r_paths);
+ _vec_len(r_paths3) = 3;
+
+ r_paths3[0].frp_weight = 3;
+
+ pl_index = fib_path_list_create(FIB_PATH_LIST_FLAG_SHARED, r_paths3);
+ fib_path_list_lock(pl_index);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[1],
+ &buckets[1],
+ &buckets[1],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0]),
+ "UCMP");
+
+ bfds[1].local_state = BFD_STATE_down;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[1]);
+
+ fib_path_list_contribute_forwarding(pl_index,
+ FIB_FORW_CHAIN_TYPE_UNICAST_IP4,
+ FIB_PATH_LIST_FWD_FLAG_STICKY,
+ &dpo);
+ /* No attempt to Un-equal distribute the down path's buckets */
+ FIB_TEST(!fib_test_validate_lb(&dpo,
+ 16,
+ &buckets[2],
+ &buckets[0],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[2],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0],
+ &buckets[0]),
+ "UCMP");
+ bfds[1].local_state = BFD_STATE_up;
+ adj_bfd_notify(BFD_LISTEN_EVENT_UPDATE, &bfds[1]);
+
+ dpo_reset(&dpo);
+ fib_path_list_unlock(pl_index);
+
+ vec_free(r_paths);
+ vec_free(r_paths2);
+ vec_free(r_paths3);
+
+ FIB_TEST(lb_count == pool_elts(load_balance_pool), "no leaked LBs");
+ FIB_TEST(pl_count == fib_path_list_pool_size(), "no leaked PLs");
+
+ return 0;
+}
+
static clib_error_t *
fib_test (vlib_main_t * vm,
unformat_input_t * input,
{
res += fib_test_inherit();
}
+ else if (unformat (input, "sticky"))
+ {
+ res += fib_test_sticky();
+ }
else
{
res += fib_test_v4();