X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fvnet%2Ftcp%2Fbuiltin_client.c;h=2784591731eb39ffd9ceee7921af39216b3b5378;hb=b384b543313b6b47a277c903e9d4fcd4343054fa;hp=27e20f8e8e50b98e15fb14425d56b7dfa0cba05b;hpb=68810624f84467503482b82662c980e8f0e36deb;p=vpp.git diff --git a/src/vnet/tcp/builtin_client.c b/src/vnet/tcp/builtin_client.c index 27e20f8e8e5..2784591731e 100644 --- a/src/vnet/tcp/builtin_client.c +++ b/src/vnet/tcp/builtin_client.c @@ -21,27 +21,9 @@ #include #include -#include #include -/* define message IDs */ -#include - -/* define message structures */ -#define vl_typedefs -#include -#undef vl_typedefs - -/* define generated endian-swappers */ -#define vl_endianfun -#include -#undef vl_endianfun - -/* instantiate all the print functions we know about */ -#define vl_print(handle, ...) vlib_cli_output (handle, __VA_ARGS__) -#define vl_printfun -#include -#undef vl_printfun +tclient_main_t tclient_main; #define TCP_BUILTIN_CLIENT_DBG (0) @@ -70,7 +52,6 @@ send_test_chunk (tclient_main_t * tm, session_t * s) int test_buf_offset; u32 bytes_this_chunk; session_fifo_event_t evt; - static int serial_number = 0; svm_fifo_t *txf; int rv; @@ -118,9 +99,8 @@ send_test_chunk (tclient_main_t * tm, session_t * s) /* Fabricate TX event, send to vpp */ evt.fifo = txf; evt.event_type = FIFO_EVENT_APP_TX; - evt.event_id = serial_number++; - if (unix_shared_memory_queue_add + if (svm_queue_add (tm->vpp_event_queue[txf->master_thread_index], (u8 *) & evt, 0 /* do wait for mutex */ )) clib_warning ("could not enqueue event"); @@ -132,13 +112,13 @@ static void receive_test_chunk (tclient_main_t * tm, session_t * s) { svm_fifo_t *rx_fifo = s->server_rx_fifo; - int n_read, test_bytes = 0; u32 my_thread_index = vlib_get_thread_index (); + int n_read, i; /* Allow enqueuing of new event */ // svm_fifo_unset_event (rx_fifo); - if (test_bytes) + if (tm->test_bytes) { n_read = svm_fifo_dequeue_nowait (rx_fifo, vec_len (tm->rx_buf[my_thread_index]), @@ -169,9 +149,8 @@ receive_test_chunk (tclient_main_t * tm, session_t * s) ed->data[0] = n_read; } - if (test_bytes) + if (tm->test_bytes) { - int i; for (i = 0; i < n_read; i++) { if (tm->rx_buf[my_thread_index][i] @@ -181,6 +160,7 @@ receive_test_chunk (tclient_main_t * tm, session_t * s) n_read, s->bytes_received + i, tm->rx_buf[my_thread_index][i], ((s->bytes_received + i) & 0xff)); + tm->test_failed = 1; } } } @@ -266,14 +246,14 @@ builtin_client_node_fn (vlib_main_t * vm, vlib_node_runtime_t * node, __sync_fetch_and_add (&tm->tx_total, sp->bytes_sent); __sync_fetch_and_add (&tm->rx_total, sp->bytes_received); - stream_session_parse_handle (sp->vpp_session_handle, - &index, &thread_index); - s = stream_session_get_if_valid (index, thread_index); + session_parse_handle (sp->vpp_session_handle, + &index, &thread_index); + s = session_get_if_valid (index, thread_index); if (s) { vnet_disconnect_args_t _a, *a = &_a; - a->handle = stream_session_handle (s); + a->handle = session_handle (s); a->app_index = tm->app_index; vnet_disconnect_session (a); @@ -308,87 +288,16 @@ VLIB_REGISTER_NODE (builtin_client_node) = }; /* *INDENT-ON* */ -/* So we don't get "no handler for... " msgs */ -static void -vl_api_memclnt_create_reply_t_handler (vl_api_memclnt_create_reply_t * mp) -{ - vlib_main_t *vm = vlib_get_main (); - tclient_main_t *tm = &tclient_main; - tm->my_client_index = mp->index; - vlib_process_signal_event (vm, tm->cli_node_index, 1 /* evt */ , - 0 /* data */ ); -} - static int create_api_loopback (tclient_main_t * tm) { - vlib_main_t *vm = vlib_get_main (); - vl_api_memclnt_create_t _m, *mp = &_m; - extern void vl_api_memclnt_create_t_handler (vl_api_memclnt_create_t *); api_main_t *am = &api_main; vl_shmem_hdr_t *shmem_hdr; - uword *event_data = 0, event_type; - int resolved = 0; - - /* - * Create a "loopback" API client connection - * Don't do things like this unless you know what you're doing... - */ shmem_hdr = am->shmem_hdr; tm->vl_input_queue = shmem_hdr->vl_input_queue; - memset (mp, 0, sizeof (*mp)); - mp->_vl_msg_id = VL_API_MEMCLNT_CREATE; - mp->context = 0xFEEDFACE; - mp->input_queue = pointer_to_uword (tm->vl_input_queue); - strncpy ((char *) mp->name, "tcp_clients_tester", sizeof (mp->name) - 1); - - vl_api_memclnt_create_t_handler (mp); - - /* Wait for reply */ - vlib_process_wait_for_event_or_clock (vm, 1.0); - event_type = vlib_process_get_events (vm, &event_data); - switch (event_type) - { - case 1: - resolved = 1; - break; - case ~0: - /* timed out */ - break; - default: - clib_warning ("unknown event_type %d", event_type); - } - if (!resolved) - return -1; - return 0; -} - -#define foreach_tclient_static_api_msg \ -_(MEMCLNT_CREATE_REPLY, memclnt_create_reply) \ - -static clib_error_t * -tclient_api_hookup (vlib_main_t * vm) -{ - vl_msg_api_msg_config_t _c, *c = &_c; - - /* Hook up client-side static APIs to our handlers */ -#define _(N,n) do { \ - c->id = VL_API_##N; \ - c->name = #n; \ - c->handler = vl_api_##n##_t_handler; \ - c->cleanup = vl_noop_handler; \ - c->endian = vl_api_##n##_t_endian; \ - c->print = vl_api_##n##_t_print; \ - c->size = sizeof(vl_api_##n##_t); \ - c->traced = 1; /* trace, so these msgs print */ \ - c->replay = 0; /* don't replay client create/delete msgs */ \ - c->message_bounce = 0; /* don't bounce this message */ \ - vl_msg_api_config(c);} while (0); - - foreach_tclient_static_api_msg; -#undef _ - + tm->my_client_index = vl_api_memclnt_create_internal ("tcp_test_client", + tm->vl_input_queue); return 0; } @@ -400,7 +309,6 @@ tcp_test_clients_init (vlib_main_t * vm) u32 num_threads; int i; - tclient_api_hookup (vm); if (create_api_loopback (tm)) return -1; @@ -433,8 +341,6 @@ builtin_session_connected_callback (u32 app_index, u32 api_context, u32 session_index; u8 thread_index = vlib_get_thread_index (); - ASSERT (s->thread_index == thread_index); - if (is_fail) { clib_warning ("connection %d failed!", api_context); @@ -442,6 +348,8 @@ builtin_session_connected_callback (u32 app_index, u32 api_context, return 0; } + ASSERT (s->thread_index == thread_index); + if (!tm->vpp_event_queue[thread_index]) tm->vpp_event_queue[thread_index] = session_manager_get_vpp_event_queue (thread_index); @@ -461,7 +369,7 @@ builtin_session_connected_callback (u32 app_index, u32 api_context, session->server_rx_fifo->client_session_index = session_index; session->server_tx_fifo = s->server_tx_fifo; session->server_tx_fifo->client_session_index = session_index; - session->vpp_session_handle = stream_session_handle (s); + session->vpp_session_handle = session_handle (s); vec_add1 (tm->connection_index_by_thread[thread_index], session_index); __sync_fetch_and_add (&tm->ready_connections, 1); @@ -495,7 +403,7 @@ builtin_session_disconnect_callback (stream_session_t * s) { tclient_main_t *tm = &tclient_main; vnet_disconnect_args_t _a, *a = &_a; - a->handle = stream_session_handle (s); + a->handle = session_handle (s); a->app_index = tm->app_index; vnet_disconnect_session (a); return; @@ -517,41 +425,45 @@ static session_cb_vft_t builtin_clients = { }; /* *INDENT-ON* */ -static int -attach_builtin_test_clients_app (void) +static clib_error_t * +attach_builtin_test_clients_app (u8 * appns_id, u64 appns_flags, + u64 appns_secret) { + u32 prealloc_fifos, segment_size = 2 << 20; tclient_main_t *tm = &tclient_main; vnet_app_attach_args_t _a, *a = &_a; - u8 segment_name[128]; - u32 segment_name_length, prealloc_fifos; u64 options[16]; - - segment_name_length = ARRAY_LEN (segment_name); + clib_error_t *error = 0; memset (a, 0, sizeof (*a)); memset (options, 0, sizeof (options)); a->api_client_index = tm->my_client_index; - a->segment_name = segment_name; - a->segment_name_length = segment_name_length; a->session_cb_vft = &builtin_clients; prealloc_fifos = tm->prealloc_fifos ? tm->expected_connections : 1; - options[SESSION_OPTIONS_ACCEPT_COOKIE] = 0x12345678; - options[SESSION_OPTIONS_SEGMENT_SIZE] = (2ULL << 32); - options[SESSION_OPTIONS_RX_FIFO_SIZE] = tm->fifo_size; - options[SESSION_OPTIONS_TX_FIFO_SIZE] = tm->fifo_size; + if (tm->private_segment_size) + segment_size = tm->private_segment_size; + + options[APP_OPTIONS_ACCEPT_COOKIE] = 0x12345678; + options[APP_OPTIONS_SEGMENT_SIZE] = segment_size; + options[APP_OPTIONS_RX_FIFO_SIZE] = tm->fifo_size; + options[APP_OPTIONS_TX_FIFO_SIZE] = tm->fifo_size; options[APP_OPTIONS_PRIVATE_SEGMENT_COUNT] = tm->private_segment_count; - options[APP_OPTIONS_PRIVATE_SEGMENT_SIZE] = tm->private_segment_size; options[APP_OPTIONS_PREALLOC_FIFO_PAIRS] = prealloc_fifos; - options[APP_OPTIONS_FLAGS] = APP_OPTIONS_FLAGS_BUILTIN_APP; - + options[APP_OPTIONS_FLAGS] = APP_OPTIONS_FLAGS_IS_BUILTIN; + if (appns_id) + { + options[APP_OPTIONS_FLAGS] |= appns_flags; + options[APP_OPTIONS_NAMESPACE_SECRET] = appns_secret; + } a->options = options; + a->namespace_id = appns_id; - if (vnet_application_attach (a)) - return -1; + if ((error = vnet_application_attach (a))) + return error; tm->app_index = a->app_index; return 0; @@ -581,11 +493,12 @@ start_tx_pthread (tclient_main_t * tm) return 0; } -void +clib_error_t * clients_connect (vlib_main_t * vm, u8 * uri, u32 n_clients) { tclient_main_t *tm = &tclient_main; vnet_connect_args_t _a, *a = &_a; + clib_error_t *error = 0; int i; for (i = 0; i < n_clients; i++) { @@ -595,13 +508,27 @@ clients_connect (vlib_main_t * vm, u8 * uri, u32 n_clients) a->api_context = i; a->app_index = tm->app_index; a->mp = 0; - vnet_connect_uri (a); - /* Crude pacing for call setups, 100k/sec */ - vlib_process_suspend (vm, 10e-6); + if ((error = vnet_connect_uri (a))) + return error; + + + /* Crude pacing for call setups */ + if ((i % 4) == 0) + vlib_process_suspend (vm, 10e-6); + ASSERT (i + 1 >= tm->ready_connections); + while (i + 1 - tm->ready_connections > 1000) + { + vlib_process_suspend (vm, 100e-6); + } } + return 0; } +#define CLI_OUTPUT(_fmt, _args...) \ + if (!tm->no_output) \ + vlib_cli_output(vm, _fmt, ##_args) + static clib_error_t * test_tcp_clients_command_fn (vlib_main_t * vm, unformat_input_t * input, @@ -610,11 +537,14 @@ test_tcp_clients_command_fn (vlib_main_t * vm, tclient_main_t *tm = &tclient_main; vlib_thread_main_t *thread_main = vlib_get_thread_main (); uword *event_data = 0, event_type; - u8 *default_connect_uri = (u8 *) "tcp://6.0.1.1/1234", *uri; - u64 tmp, total_bytes; - f64 cli_timeout = 20.0, delta; + u8 *default_connect_uri = (u8 *) "tcp://6.0.1.1/1234", *uri, *appns_id = 0; + u64 tmp, total_bytes, appns_flags = 0, appns_secret = 0; + f64 test_timeout = 20.0, syn_timeout = 20.0, delta; + f64 time_before_connects; u32 n_clients = 1; + int preallocate_sessions = 0; char *transfer_type; + clib_error_t *error = 0; int i; tm->bytes_to_send = 8192; @@ -623,6 +553,9 @@ test_tcp_clients_command_fn (vlib_main_t * vm, tm->connections_per_batch = 1000; tm->private_segment_count = 0; tm->private_segment_size = 0; + tm->no_output = 0; + tm->test_bytes = 0; + tm->test_failed = 0; tm->vlib_main = vm; if (thread_main->n_vlib_mains > 1) clib_spinlock_init (&tm->sessions_lock); @@ -640,7 +573,9 @@ test_tcp_clients_command_fn (vlib_main_t * vm, ; else if (unformat (input, "uri %s", &tm->connect_uri)) ; - else if (unformat (input, "cli-timeout %f", &cli_timeout)) + else if (unformat (input, "test-timeout %f", &test_timeout)) + ; + else if (unformat (input, "syn-timeout %f", &syn_timeout)) ; else if (unformat (input, "no-return")) tm->no_return = 1; @@ -649,17 +584,36 @@ test_tcp_clients_command_fn (vlib_main_t * vm, else if (unformat (input, "private-segment-count %d", &tm->private_segment_count)) ; - else if (unformat (input, "private-segment-size %dm", &tmp)) - tm->private_segment_size = tmp << 20; - else if (unformat (input, "private-segment-size %dg", &tmp)) - tm->private_segment_size = tmp << 30; - else if (unformat (input, "private-segment-size %d", &tmp)) - tm->private_segment_size = tmp; + else if (unformat (input, "private-segment-size %U", + unformat_memory_size, &tmp)) + { + if (tmp >= 0x100000000ULL) + return clib_error_return + (0, "private segment size %lld (%llu) too large", tmp, tmp); + tm->private_segment_size = tmp; + } else if (unformat (input, "preallocate-fifos")) tm->prealloc_fifos = 1; + else if (unformat (input, "preallocate-sessions")) + preallocate_sessions = 1; else if (unformat (input, "client-batch %d", &tm->connections_per_batch)) ; + else if (unformat (input, "appns %_%v%_", &appns_id)) + ; + else if (unformat (input, "all-scope")) + appns_flags |= (APP_OPTIONS_FLAGS_USE_GLOBAL_SCOPE + | APP_OPTIONS_FLAGS_USE_LOCAL_SCOPE); + else if (unformat (input, "local-scope")) + appns_flags = APP_OPTIONS_FLAGS_USE_LOCAL_SCOPE; + else if (unformat (input, "global-scope")) + appns_flags = APP_OPTIONS_FLAGS_USE_GLOBAL_SCOPE; + else if (unformat (input, "secret %lu", &appns_secret)) + ; + else if (unformat (input, "no-output")) + tm->no_output = 1; + else if (unformat (input, "test-bytes")) + tm->test_bytes = 1; else return clib_error_return (0, "unknown input `%U'", format_unformat_error, input); @@ -674,6 +628,7 @@ test_tcp_clients_command_fn (vlib_main_t * vm, return clib_error_return (0, "failed init"); } + tm->ready_connections = 0; tm->expected_connections = n_clients; tm->rx_total = 0; @@ -693,10 +648,14 @@ test_tcp_clients_command_fn (vlib_main_t * vm, if (tm->test_client_attached == 0) { - if (attach_builtin_test_clients_app ()) + if ((error = attach_builtin_test_clients_app (appns_id, appns_flags, + appns_secret))) { - return clib_error_return (0, "app attach failed"); + vec_free (appns_id); + clib_error_report (error); + return error; } + vec_free (appns_id); } tm->test_client_attached = 1; @@ -705,46 +664,70 @@ test_tcp_clients_command_fn (vlib_main_t * vm, vlib_node_set_state (vlib_mains[i], builtin_client_node.index, VLIB_NODE_STATE_POLLING); + if (preallocate_sessions) + { + session_t *sp __attribute__ ((unused)); + for (i = 0; i < n_clients; i++) + pool_get (tm->sessions, sp); + for (i = 0; i < n_clients; i++) + pool_put_index (tm->sessions, i); + } + /* Fire off connect requests */ - clients_connect (vm, uri, n_clients); + time_before_connects = vlib_time_now (vm); + if ((error = clients_connect (vm, uri, n_clients))) + return error; /* Park until the sessions come up, or ten seconds elapse... */ - vlib_process_wait_for_event_or_clock (vm, 10 /* timeout, seconds */ ); + vlib_process_wait_for_event_or_clock (vm, syn_timeout); event_type = vlib_process_get_events (vm, &event_data); switch (event_type) { case ~0: - vlib_cli_output (vm, "Timeout with only %d sessions active...", - tm->ready_connections); + CLI_OUTPUT ("Timeout with only %d sessions active...", + tm->ready_connections); + error = clib_error_return (0, "failed: syn timeout with %d sessions", + tm->ready_connections); goto cleanup; case 1: + delta = vlib_time_now (vm) - time_before_connects; + if (delta != 0.0) + CLI_OUTPUT ("%d three-way handshakes in %.2f seconds %.2f/s", + n_clients, delta, ((f64) n_clients) / delta); + tm->test_start_time = vlib_time_now (tm->vlib_main); - vlib_cli_output (vm, "Test started at %.6f", tm->test_start_time); + CLI_OUTPUT ("Test started at %.6f", tm->test_start_time); break; default: - vlib_cli_output (vm, "unexpected event(1): %d", event_type); + CLI_OUTPUT ("unexpected event(1): %d", event_type); + error = clib_error_return (0, "failed: unexpected event(1): %d", + event_type); goto cleanup; } /* Now wait for the sessions to finish... */ - vlib_process_wait_for_event_or_clock (vm, cli_timeout); + vlib_process_wait_for_event_or_clock (vm, test_timeout); event_type = vlib_process_get_events (vm, &event_data); switch (event_type) { case ~0: - vlib_cli_output (vm, "Timeout with %d sessions still active...", - tm->ready_connections); + CLI_OUTPUT ("Timeout with %d sessions still active...", + tm->ready_connections); + error = clib_error_return (0, "failed: timeout with %d sessions", + tm->ready_connections); goto cleanup; case 2: tm->test_end_time = vlib_time_now (vm); - vlib_cli_output (vm, "Test finished at %.6f", tm->test_end_time); + CLI_OUTPUT ("Test finished at %.6f", tm->test_end_time); break; default: - vlib_cli_output (vm, "unexpected event(2): %d", event_type); + CLI_OUTPUT ("unexpected event(2): %d", event_type); + error = clib_error_return (0, "failed: unexpected event(2): %d", + event_type); goto cleanup; } @@ -754,18 +737,23 @@ test_tcp_clients_command_fn (vlib_main_t * vm, { total_bytes = (tm->no_return ? tm->tx_total : tm->rx_total); transfer_type = tm->no_return ? "half-duplex" : "full-duplex"; - vlib_cli_output (vm, - "%lld bytes (%lld mbytes, %lld gbytes) in %.2f seconds", - total_bytes, total_bytes / (1ULL << 20), - total_bytes / (1ULL << 30), delta); - vlib_cli_output (vm, "%.2f bytes/second %s", - ((f64) total_bytes) / (delta), transfer_type); - vlib_cli_output (vm, "%.4f gbit/second %s", - (((f64) total_bytes * 8.0) / delta / 1e9), - transfer_type); + CLI_OUTPUT ("%lld bytes (%lld mbytes, %lld gbytes) in %.2f seconds", + total_bytes, total_bytes / (1ULL << 20), + total_bytes / (1ULL << 30), delta); + CLI_OUTPUT ("%.2f bytes/second %s", ((f64) total_bytes) / (delta), + transfer_type); + CLI_OUTPUT ("%.4f gbit/second %s", + (((f64) total_bytes * 8.0) / delta / 1e9), transfer_type); } else - vlib_cli_output (vm, "zero delta-t?"); + { + CLI_OUTPUT ("zero delta-t?"); + error = clib_error_return (0, "failed: zero delta-t"); + goto cleanup; + } + + if (tm->test_bytes && tm->test_failed) + error = clib_error_return (0, "failed: test bytes"); cleanup: tm->run_test = 0; @@ -777,15 +765,36 @@ cleanup: pool_free (tm->sessions); - return 0; + /* Detach the application, so we can use different fifo sizes next time */ + if (tm->test_client_attached) + { + vnet_app_detach_args_t _da, *da = &_da; + int rv; + + da->app_index = tm->app_index; + rv = vnet_application_detach (da); + if (rv) + { + error = clib_error_return (0, "failed: app detach"); + CLI_OUTPUT ("WARNING: app detach failed..."); + } + tm->test_client_attached = 0; + tm->app_index = ~0; + } + if (error) + CLI_OUTPUT ("test failed"); + return error; } /* *INDENT-OFF* */ VLIB_CLI_COMMAND (test_clients_command, static) = { .path = "test tcp clients", - .short_help = "test tcp clients [nclients %d]" - "[iterations %d] [bytes %d] [uri tcp://6.0.1.1/1234]", + .short_help = "test tcp clients [nclients %d] [[m|g]bytes ] " + "[test-timeout