X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Ftcp%2Ftcp_cubic.c;h=b79ef8342d30f7a0b1fd1e7b342df649ff85f3eb;hb=282872127;hp=b9a1c3da06a8b3dfd7092f3e66455e3eba12886f;hpb=2e31cc35ca5db7f16c8052578d79f1ec84c0acb5;p=vpp.git diff --git a/src/vnet/tcp/tcp_cubic.c b/src/vnet/tcp/tcp_cubic.c index b9a1c3da06a..b79ef8342d3 100644 --- a/src/vnet/tcp/tcp_cubic.c +++ b/src/vnet/tcp/tcp_cubic.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2018 Cisco and/or its affiliates. + * Copyright (c) 2018-2019 Cisco and/or its affiliates. * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at: @@ -20,6 +20,17 @@ #define cubic_c 0.4 #define west_const (3 * (1 - beta_cubic) / (1 + beta_cubic)) +typedef struct cubic_cfg_ +{ + u8 fast_convergence; + u32 ssthresh; +} cubic_cfg_t; + +static cubic_cfg_t cubic_cfg = { + .fast_convergence = 1, + .ssthresh = 0x7FFFFFFFU, +}; + typedef struct cubic_data_ { /** time period (in seconds) needed to increase the current window @@ -29,7 +40,7 @@ typedef struct cubic_data_ /** time (in sec) since the start of current congestion avoidance */ f64 t_start; - /** Inflection point of the cubic function */ + /** Inflection point of the cubic function (in snd_mss segments) */ u32 w_max; } __clib_packed cubic_data_t; @@ -60,10 +71,12 @@ W_cubic (cubic_data_t * cd, f64 t) * RFC 8312 Eq. 2 */ static inline f64 -K_cubic (cubic_data_t * cd) +K_cubic (cubic_data_t * cd, u32 wnd) { - /* K = cubic_root(W_max*(1-beta_cubic)/C) */ - return pow (cd->w_max * (1 - beta_cubic) / cubic_c, 1 / 3.0); + /* K = cubic_root(W_max*(1-beta_cubic)/C) + * Because the current window may be less than W_max * beta_cubic because + * of fast convergence, we pass it as parameter */ + return pow ((f64) (cd->w_max - wnd) / cubic_c, 1 / 3.0); } /** @@ -84,9 +97,26 @@ static void cubic_congestion (tcp_connection_t * tc) { cubic_data_t *cd = (cubic_data_t *) tcp_cc_data (tc); + u32 w_max; - cd->w_max = tc->cwnd / tc->snd_mss; + w_max = tc->cwnd / tc->snd_mss; + if (cubic_cfg.fast_convergence && w_max < cd->w_max) + w_max = w_max * ((1.0 + beta_cubic) / 2.0); + + cd->w_max = w_max; tc->ssthresh = clib_max (tc->cwnd * beta_cubic, 2 * tc->snd_mss); + tc->cwnd = tc->ssthresh; +} + +static void +cubic_loss (tcp_connection_t * tc) +{ + cubic_data_t *cd = (cubic_data_t *) tcp_cc_data (tc); + + tc->cwnd = tcp_loss_wnd (tc); + cd->t_start = cubic_time (tc->c_thread_index); + cd->K = 0; + cd->w_max = tc->cwnd / tc->snd_mss; } static void @@ -94,12 +124,27 @@ cubic_recovered (tcp_connection_t * tc) { cubic_data_t *cd = (cubic_data_t *) tcp_cc_data (tc); cd->t_start = cubic_time (tc->c_thread_index); - cd->K = K_cubic (cd); tc->cwnd = tc->ssthresh; + cd->K = K_cubic (cd, tc->cwnd / tc->snd_mss); } static void -cubic_rcv_ack (tcp_connection_t * tc) +cubic_cwnd_accumulate (tcp_connection_t * tc, u32 thresh, u32 bytes_acked) +{ + /* We just updated the threshold and don't know how large the previous + * one was. Still, optimistically increase cwnd by one segment and + * clear the accumulated bytes. */ + if (tc->cwnd_acc_bytes > thresh) + { + tc->cwnd += tc->snd_mss; + tc->cwnd_acc_bytes = 0; + } + + tcp_cwnd_accumulate (tc, thresh, tc->bytes_acked); +} + +static void +cubic_rcv_ack (tcp_connection_t * tc, tcp_rate_sample_t * rs) { cubic_data_t *cd = (cubic_data_t *) tcp_cc_data (tc); u64 w_cubic, w_aimd; @@ -112,7 +157,7 @@ cubic_rcv_ack (tcp_connection_t * tc) if (tcp_in_slowstart (tc)) { - tc->cwnd += clib_min (tc->snd_mss, tc->bytes_acked); + tc->cwnd += tc->bytes_acked; return; } @@ -120,10 +165,10 @@ cubic_rcv_ack (tcp_connection_t * tc) rtt_sec = clib_min (tc->mrtt_us, (f64) tc->srtt * TCP_TICK); w_cubic = W_cubic (cd, t + rtt_sec) * tc->snd_mss; - w_aimd = W_est (cd, t, rtt_sec) * tc->snd_mss; + w_aimd = (u64) W_est (cd, t, rtt_sec) * tc->snd_mss; if (w_cubic < w_aimd) { - tcp_cwnd_accumulate (tc, tc->cwnd, tc->bytes_acked); + cubic_cwnd_accumulate (tc, tc->cwnd, tc->bytes_acked); } else { @@ -141,16 +186,15 @@ cubic_rcv_ack (tcp_connection_t * tc) */ thresh = (tc->snd_mss * tc->cwnd) / (w_cubic - tc->cwnd); - /* Make sure we don't increase cwnd more often than every - * 2 segments */ - thresh = clib_max (thresh, 2 * tc->snd_mss); + /* Make sure we don't increase cwnd more often than every segment */ + thresh = clib_max (thresh, tc->snd_mss); } else { /* Practically we can't increment so just inflate threshold */ - thresh = 1000 * tc->cwnd; + thresh = 50 * tc->cwnd; } - tcp_cwnd_accumulate (tc, thresh, tc->bytes_acked); + cubic_cwnd_accumulate (tc, thresh, tc->bytes_acked); } } @@ -158,19 +202,44 @@ static void cubic_conn_init (tcp_connection_t * tc) { cubic_data_t *cd = (cubic_data_t *) tcp_cc_data (tc); - tc->ssthresh = tc->snd_wnd; + tc->ssthresh = cubic_cfg.ssthresh; tc->cwnd = tcp_initial_cwnd (tc); cd->w_max = 0; cd->K = 0; cd->t_start = cubic_time (tc->c_thread_index); } +static uword +cubic_unformat_config (unformat_input_t * input) +{ + u32 ssthresh = 0x7FFFFFFFU; + + if (!input) + return 0; + + unformat_skip_white_space (input); + + while (unformat_check_input (input) != UNFORMAT_END_OF_INPUT) + { + if (unformat (input, "no-fast-convergence")) + cubic_cfg.fast_convergence = 0; + else if (unformat (input, "ssthresh %u", &ssthresh)) + cubic_cfg.ssthresh = ssthresh; + else + return 0; + } + return 1; +} + const static tcp_cc_algorithm_t tcp_cubic = { + .name = "cubic", + .unformat_cfg = cubic_unformat_config, .congestion = cubic_congestion, + .loss = cubic_loss, .recovered = cubic_recovered, .rcv_ack = cubic_rcv_ack, .rcv_cong_ack = newreno_rcv_cong_ack, - .init = cubic_conn_init + .init = cubic_conn_init, }; clib_error_t *