2 * Copyright (c) 2015 Cisco and/or its affiliates.
3 * Licensed under the Apache License, Version 2.0 (the "License");
4 * you may not use this file except in compliance with the License.
5 * You may obtain a copy of the License at:
7 * http://www.apache.org/licenses/LICENSE-2.0
9 * Unless required by applicable law or agreed to in writing, software
10 * distributed under the License is distributed on an "AS IS" BASIS,
11 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 * See the License for the specific language governing permissions and
13 * limitations under the License.
16 * pg_input.c: buffer generator input
18 * Copyright (c) 2008 Eliot Dresselhaus
20 * Permission is hereby granted, free of charge, to any person obtaining
21 * a copy of this software and associated documentation files (the
22 * "Software"), to deal in the Software without restriction, including
23 * without limitation the rights to use, copy, modify, merge, publish,
24 * distribute, sublicense, and/or sell copies of the Software, and to
25 * permit persons to whom the Software is furnished to do so, subject to
26 * the following conditions:
28 * The above copyright notice and this permission notice shall be
29 * included in all copies or substantial portions of the Software.
31 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
32 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
33 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
34 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
35 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
36 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
37 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
41 * To be honest, the packet generator needs an extreme
42 * makeover. Two key assumptions which drove the current implementation
43 * are no longer true. First, buffer managers implement a
44 * post-TX recycle list. Second, that packet generator performance
45 * is first-order important.
48 #include <vlib/vlib.h>
49 #include <vnet/pg/pg.h>
50 #include <vnet/vnet.h>
51 #include <vnet/ethernet/ethernet.h>
52 #include <vnet/feature/feature.h>
53 #include <vnet/ip/ip4_packet.h>
54 #include <vnet/ip/ip6_packet.h>
55 #include <vnet/udp/udp_packet.h>
56 #include <vnet/devices/devices.h>
57 #include <vnet/gso/gro_func.h>
60 validate_buffer_data2 (vlib_buffer_t * b, pg_stream_t * s,
61 u32 data_offset, u32 n_bytes)
67 pd = s->fixed_packet_data + data_offset;
68 pm = s->fixed_packet_data_mask + data_offset;
70 if (pd + n_bytes >= vec_end (s->fixed_packet_data))
71 n_bytes = (pd < vec_end (s->fixed_packet_data)
72 ? vec_end (s->fixed_packet_data) - pd : 0);
74 for (i = 0; i < n_bytes; i++)
75 if ((bd[i] & pm[i]) != pd[i])
81 clib_warning ("buffer %U", format_vnet_buffer, b);
82 clib_warning ("differ at index %d", i);
83 clib_warning ("is %U", format_hex_bytes, bd, n_bytes);
84 clib_warning ("mask %U", format_hex_bytes, pm, n_bytes);
85 clib_warning ("expect %U", format_hex_bytes, pd, n_bytes);
90 validate_buffer_data (vlib_buffer_t * b, pg_stream_t * s)
92 return validate_buffer_data2 (b, s, 0, s->buffer_bytes);
97 u64 v0, u64 v_min, u64 v_max, u32 n_bits, u32 is_net_byte_order)
99 ASSERT (v0 >= v_min && v0 <= v_max);
100 if (n_bits == BITS (u8))
104 else if (n_bits == BITS (u16))
106 if (is_net_byte_order)
107 v0 = clib_host_to_net_u16 (v0);
108 clib_mem_unaligned (a0, u16) = v0;
110 else if (n_bits == BITS (u32))
112 if (is_net_byte_order)
113 v0 = clib_host_to_net_u32 (v0);
114 clib_mem_unaligned (a0, u32) = v0;
116 else if (n_bits == BITS (u64))
118 if (is_net_byte_order)
119 v0 = clib_host_to_net_u64 (v0);
120 clib_mem_unaligned (a0, u64) = v0;
125 set_2 (void *a0, void *a1,
127 u64 v_min, u64 v_max,
128 u32 n_bits, u32 is_net_byte_order, u32 is_increment)
130 ASSERT (v0 >= v_min && v0 <= v_max);
131 ASSERT (v1 >= v_min && v1 <= (v_max + is_increment));
132 if (n_bits == BITS (u8))
137 else if (n_bits == BITS (u16))
139 if (is_net_byte_order)
141 v0 = clib_host_to_net_u16 (v0);
142 v1 = clib_host_to_net_u16 (v1);
144 clib_mem_unaligned (a0, u16) = v0;
145 clib_mem_unaligned (a1, u16) = v1;
147 else if (n_bits == BITS (u32))
149 if (is_net_byte_order)
151 v0 = clib_host_to_net_u32 (v0);
152 v1 = clib_host_to_net_u32 (v1);
154 clib_mem_unaligned (a0, u32) = v0;
155 clib_mem_unaligned (a1, u32) = v1;
157 else if (n_bits == BITS (u64))
159 if (is_net_byte_order)
161 v0 = clib_host_to_net_u64 (v0);
162 v1 = clib_host_to_net_u64 (v1);
164 clib_mem_unaligned (a0, u64) = v0;
165 clib_mem_unaligned (a1, u64) = v1;
169 static_always_inline void
170 do_set_fixed (pg_main_t * pg,
175 u32 byte_offset, u32 is_net_byte_order, u64 v_min, u64 v_max)
177 vlib_main_t *vm = vlib_get_main ();
179 while (n_buffers >= 4)
181 vlib_buffer_t *b0, *b1, *b2, *b3;
184 b0 = vlib_get_buffer (vm, buffers[0]);
185 b1 = vlib_get_buffer (vm, buffers[1]);
186 b2 = vlib_get_buffer (vm, buffers[2]);
187 b3 = vlib_get_buffer (vm, buffers[3]);
191 a0 = (void *) b0 + byte_offset;
192 a1 = (void *) b1 + byte_offset;
193 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
194 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
196 set_2 (a0, a1, v_min, v_min, v_min, v_max, n_bits, is_net_byte_order,
197 /* is_increment */ 0);
199 ASSERT (validate_buffer_data (b0, s));
200 ASSERT (validate_buffer_data (b1, s));
203 while (n_buffers > 0)
208 b0 = vlib_get_buffer (vm, buffers[0]);
212 a0 = (void *) b0 + byte_offset;
214 set_1 (a0, v_min, v_min, v_max, n_bits, is_net_byte_order);
216 ASSERT (validate_buffer_data (b0, s));
220 static_always_inline u64
221 do_set_increment (pg_main_t * pg,
227 u32 is_net_byte_order,
228 u32 want_sum, u64 * sum_result, u64 v_min, u64 v_max, u64 v)
230 vlib_main_t *vm = vlib_get_main ();
233 ASSERT (v >= v_min && v <= v_max);
235 while (n_buffers >= 4)
237 vlib_buffer_t *b0, *b1, *b2, *b3;
241 b0 = vlib_get_buffer (vm, buffers[0]);
242 b1 = vlib_get_buffer (vm, buffers[1]);
243 b2 = vlib_get_buffer (vm, buffers[2]);
244 b3 = vlib_get_buffer (vm, buffers[3]);
248 a0 = (void *) b0 + byte_offset;
249 a1 = (void *) b1 + byte_offset;
250 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
251 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
255 v = v > v_max ? v_min : v;
257 v_old + 0, v_old + 1, v_min, v_max, n_bits, is_net_byte_order,
258 /* is_increment */ 1);
261 sum += 2 * v_old + 1;
263 if (PREDICT_FALSE (v_old + 1 > v_max))
266 sum -= 2 * v_old + 1;
269 set_1 (a0, v + 0, v_min, v_max, n_bits, is_net_byte_order);
274 v = v > v_max ? v_min : v;
275 set_1 (a1, v + 0, v_min, v_max, n_bits, is_net_byte_order);
281 ASSERT (validate_buffer_data (b0, s));
282 ASSERT (validate_buffer_data (b1, s));
285 while (n_buffers > 0)
291 b0 = vlib_get_buffer (vm, buffers[0]);
295 a0 = (void *) b0 + byte_offset;
301 v = v > v_max ? v_min : v;
303 ASSERT (v_old >= v_min && v_old <= v_max);
304 set_1 (a0, v_old, v_min, v_max, n_bits, is_net_byte_order);
306 ASSERT (validate_buffer_data (b0, s));
315 static_always_inline void
316 do_set_random (pg_main_t * pg,
322 u32 is_net_byte_order,
323 u32 want_sum, u64 * sum_result, u64 v_min, u64 v_max)
325 vlib_main_t *vm = vlib_get_main ();
326 u64 v_diff = v_max - v_min + 1;
327 u64 r_mask = max_pow2 (v_diff) - 1;
332 random_data = clib_random_buffer_get_data
333 (&vm->random_buffer, n_buffers * n_bits / BITS (u8));
337 while (n_buffers >= 4)
339 vlib_buffer_t *b0, *b1, *b2, *b3;
341 u64 r0 = 0, r1 = 0; /* warnings be gone */
343 b0 = vlib_get_buffer (vm, buffers[0]);
344 b1 = vlib_get_buffer (vm, buffers[1]);
345 b2 = vlib_get_buffer (vm, buffers[2]);
346 b3 = vlib_get_buffer (vm, buffers[3]);
350 a0 = (void *) b0 + byte_offset;
351 a1 = (void *) b1 + byte_offset;
352 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
353 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
360 u##n * r = random_data; \
363 random_data = r + 2; \
375 /* Add power of 2 sized random number which may be out of range. */
379 /* Twice should be enough to reduce to v_min .. v_max range. */
380 v0 = v0 > v_max ? v0 - v_diff : v0;
381 v1 = v1 > v_max ? v1 - v_diff : v1;
382 v0 = v0 > v_max ? v0 - v_diff : v0;
383 v1 = v1 > v_max ? v1 - v_diff : v1;
388 set_2 (a0, a1, v0, v1, v_min, v_max, n_bits, is_net_byte_order,
389 /* is_increment */ 0);
391 ASSERT (validate_buffer_data (b0, s));
392 ASSERT (validate_buffer_data (b1, s));
395 while (n_buffers > 0)
399 u64 r0 = 0; /* warnings be gone */
401 b0 = vlib_get_buffer (vm, buffers[0]);
405 a0 = (void *) b0 + byte_offset;
412 u##n * r = random_data; \
414 random_data = r + 1; \
426 /* Add power of 2 sized random number which may be out of range. */
429 /* Twice should be enough to reduce to v_min .. v_max range. */
430 v0 = v0 > v_max ? v0 - v_diff : v0;
431 v0 = v0 > v_max ? v0 - v_diff : v0;
436 set_1 (a0, v0, v_min, v_max, n_bits, is_net_byte_order);
438 ASSERT (validate_buffer_data (b0, s));
446 clib_mem_unaligned (a##i, t) = \
447 clib_host_to_net_##t ((clib_net_to_host_mem_##t (a##i) &~ mask) \
453 u64 v_min, u64 v_max,
454 u32 max_bits, u32 n_bits, u64 mask, u32 shift)
456 ASSERT (v0 >= v_min && v0 <= v_max);
457 if (max_bits == BITS (u8))
458 ((u8 *) a0)[0] = (((u8 *) a0)[0] & ~mask) | (v0 << shift);
460 else if (max_bits == BITS (u16))
464 else if (max_bits == BITS (u32))
468 else if (max_bits == BITS (u64))
475 setbits_2 (void *a0, void *a1,
477 u64 v_min, u64 v_max,
478 u32 max_bits, u32 n_bits, u64 mask, u32 shift, u32 is_increment)
480 ASSERT (v0 >= v_min && v0 <= v_max);
481 ASSERT (v1 >= v_min && v1 <= v_max + is_increment);
482 if (max_bits == BITS (u8))
484 ((u8 *) a0)[0] = (((u8 *) a0)[0] & ~mask) | (v0 << shift);
485 ((u8 *) a1)[0] = (((u8 *) a1)[0] & ~mask) | (v1 << shift);
488 else if (max_bits == BITS (u16))
493 else if (max_bits == BITS (u32))
498 else if (max_bits == BITS (u64))
507 static_always_inline void
508 do_setbits_fixed (pg_main_t * pg,
514 u32 byte_offset, u64 v_min, u64 v_max, u64 mask, u32 shift)
516 vlib_main_t *vm = vlib_get_main ();
518 while (n_buffers >= 4)
520 vlib_buffer_t *b0, *b1, *b2, *b3;
523 b0 = vlib_get_buffer (vm, buffers[0]);
524 b1 = vlib_get_buffer (vm, buffers[1]);
525 b2 = vlib_get_buffer (vm, buffers[2]);
526 b3 = vlib_get_buffer (vm, buffers[3]);
530 a0 = (void *) b0 + byte_offset;
531 a1 = (void *) b1 + byte_offset;
532 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
533 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
536 v_min, v_min, v_min, v_max, max_bits, n_bits, mask, shift,
537 /* is_increment */ 0);
539 ASSERT (validate_buffer_data (b0, s));
540 ASSERT (validate_buffer_data (b1, s));
543 while (n_buffers > 0)
548 b0 = vlib_get_buffer (vm, buffers[0]);
552 a0 = (void *) b0 + byte_offset;
554 setbits_1 (a0, v_min, v_min, v_max, max_bits, n_bits, mask, shift);
555 ASSERT (validate_buffer_data (b0, s));
559 static_always_inline u64
560 do_setbits_increment (pg_main_t * pg,
567 u64 v_min, u64 v_max, u64 v, u64 mask, u32 shift)
569 vlib_main_t *vm = vlib_get_main ();
571 ASSERT (v >= v_min && v <= v_max);
573 while (n_buffers >= 4)
575 vlib_buffer_t *b0, *b1, *b2, *b3;
579 b0 = vlib_get_buffer (vm, buffers[0]);
580 b1 = vlib_get_buffer (vm, buffers[1]);
581 b2 = vlib_get_buffer (vm, buffers[2]);
582 b3 = vlib_get_buffer (vm, buffers[3]);
586 a0 = (void *) b0 + byte_offset;
587 a1 = (void *) b1 + byte_offset;
588 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
589 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
593 v = v > v_max ? v_min : v;
595 v_old + 0, v_old + 1,
596 v_min, v_max, max_bits, n_bits, mask, shift,
597 /* is_increment */ 1);
599 if (PREDICT_FALSE (v_old + 1 > v_max))
602 setbits_1 (a0, v + 0, v_min, v_max, max_bits, n_bits, mask, shift);
605 v = v > v_max ? v_min : v;
606 setbits_1 (a1, v + 0, v_min, v_max, max_bits, n_bits, mask, shift);
609 ASSERT (validate_buffer_data (b0, s));
610 ASSERT (validate_buffer_data (b1, s));
613 while (n_buffers > 0)
619 b0 = vlib_get_buffer (vm, buffers[0]);
623 a0 = (void *) b0 + byte_offset;
627 v = v > v_max ? v_min : v;
629 ASSERT (v_old >= v_min && v_old <= v_max);
630 setbits_1 (a0, v_old, v_min, v_max, max_bits, n_bits, mask, shift);
632 ASSERT (validate_buffer_data (b0, s));
638 static_always_inline void
639 do_setbits_random (pg_main_t * pg,
645 u32 byte_offset, u64 v_min, u64 v_max, u64 mask, u32 shift)
647 vlib_main_t *vm = vlib_get_main ();
648 u64 v_diff = v_max - v_min + 1;
649 u64 r_mask = max_pow2 (v_diff) - 1;
653 random_data = clib_random_buffer_get_data
654 (&vm->random_buffer, n_buffers * max_bits / BITS (u8));
657 while (n_buffers >= 4)
659 vlib_buffer_t *b0, *b1, *b2, *b3;
661 u64 r0 = 0, r1 = 0; /* warnings be gone */
663 b0 = vlib_get_buffer (vm, buffers[0]);
664 b1 = vlib_get_buffer (vm, buffers[1]);
665 b2 = vlib_get_buffer (vm, buffers[2]);
666 b3 = vlib_get_buffer (vm, buffers[3]);
670 a0 = (void *) b0 + byte_offset;
671 a1 = (void *) b1 + byte_offset;
672 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
673 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
680 u##n * r = random_data; \
683 random_data = r + 2; \
695 /* Add power of 2 sized random number which may be out of range. */
699 /* Twice should be enough to reduce to v_min .. v_max range. */
700 v0 = v0 > v_max ? v0 - v_diff : v0;
701 v1 = v1 > v_max ? v1 - v_diff : v1;
702 v0 = v0 > v_max ? v0 - v_diff : v0;
703 v1 = v1 > v_max ? v1 - v_diff : v1;
705 setbits_2 (a0, a1, v0, v1, v_min, v_max, max_bits, n_bits, mask, shift,
706 /* is_increment */ 0);
708 ASSERT (validate_buffer_data (b0, s));
709 ASSERT (validate_buffer_data (b1, s));
712 while (n_buffers > 0)
716 u64 r0 = 0; /* warnings be gone */
718 b0 = vlib_get_buffer (vm, buffers[0]);
722 a0 = (void *) b0 + byte_offset;
729 u##n * r = random_data; \
731 random_data = r + 1; \
743 /* Add power of 2 sized random number which may be out of range. */
746 /* Twice should be enough to reduce to v_min .. v_max range. */
747 v0 = v0 > v_max ? v0 - v_diff : v0;
748 v0 = v0 > v_max ? v0 - v_diff : v0;
750 setbits_1 (a0, v0, v_min, v_max, max_bits, n_bits, mask, shift);
752 ASSERT (validate_buffer_data (b0, s));
757 do_it (pg_main_t * pg,
761 u32 lo_bit, u32 hi_bit,
762 u64 v_min, u64 v_max, u64 v, pg_edit_type_t edit_type)
764 u32 max_bits, l0, l1, h1, start_bit;
767 edit_type = PG_EDIT_FIXED;
769 l0 = lo_bit / BITS (u8);
770 l1 = lo_bit % BITS (u8);
771 h1 = hi_bit % BITS (u8);
773 start_bit = l0 * BITS (u8);
775 max_bits = hi_bit - start_bit;
776 ASSERT (max_bits <= 64);
780 if (edit_type == PG_EDIT_INCREMENT) \
781 v = do_set_increment (pg, s, buffers, n_buffers, \
784 /* is_net_byte_order */ 1, \
785 /* want sum */ 0, 0, \
788 else if (edit_type == PG_EDIT_RANDOM) \
789 do_set_random (pg, s, buffers, n_buffers, \
792 /* is_net_byte_order */ 1, \
793 /* want sum */ 0, 0, \
795 else /* edit_type == PG_EDIT_FIXED */ \
796 do_set_fixed (pg, s, buffers, n_buffers, \
799 /* is_net_byte_order */ 1, \
803 if (l1 == 0 && h1 == 0)
819 u32 n_bits = max_bits;
821 max_bits = clib_max (max_pow2 (n_bits), 8);
823 mask = ((u64) 1 << (u64) n_bits) - 1;
824 mask &= ~(((u64) 1 << (u64) shift) - 1);
826 mask <<= max_bits - n_bits;
827 shift += max_bits - n_bits;
833 if (edit_type == PG_EDIT_INCREMENT) \
834 v = do_setbits_increment (pg, s, buffers, n_buffers, \
835 BITS (u##n), n_bits, \
836 l0, v_min, v_max, v, \
838 else if (edit_type == PG_EDIT_RANDOM) \
839 do_setbits_random (pg, s, buffers, n_buffers, \
840 BITS (u##n), n_bits, \
843 else /* edit_type == PG_EDIT_FIXED */ \
844 do_setbits_fixed (pg, s, buffers, n_buffers, \
845 BITS (u##n), n_bits, \
864 pg_generate_set_lengths (pg_main_t * pg,
865 pg_stream_t * s, u32 * buffers, u32 n_buffers)
867 u64 v_min, v_max, length_sum;
868 pg_edit_type_t edit_type;
870 v_min = s->min_packet_bytes;
871 v_max = s->max_packet_bytes;
872 edit_type = s->packet_size_edit_type;
874 if (edit_type == PG_EDIT_INCREMENT)
875 s->last_increment_packet_size
876 = do_set_increment (pg, s, buffers, n_buffers,
877 8 * STRUCT_SIZE_OF (vlib_buffer_t, current_length),
878 STRUCT_OFFSET_OF (vlib_buffer_t, current_length),
879 /* is_net_byte_order */ 0,
880 /* want sum */ 1, &length_sum,
881 v_min, v_max, s->last_increment_packet_size);
883 else if (edit_type == PG_EDIT_RANDOM)
884 do_set_random (pg, s, buffers, n_buffers,
885 8 * STRUCT_SIZE_OF (vlib_buffer_t, current_length),
886 STRUCT_OFFSET_OF (vlib_buffer_t, current_length),
887 /* is_net_byte_order */ 0,
888 /* want sum */ 1, &length_sum,
891 else /* edit_type == PG_EDIT_FIXED */
893 do_set_fixed (pg, s, buffers, n_buffers,
894 8 * STRUCT_SIZE_OF (vlib_buffer_t, current_length),
895 STRUCT_OFFSET_OF (vlib_buffer_t, current_length),
896 /* is_net_byte_order */ 0,
898 length_sum = v_min * n_buffers;
902 vnet_main_t *vnm = vnet_get_main ();
903 vnet_interface_main_t *im = &vnm->interface_main;
904 vnet_sw_interface_t *si =
905 vnet_get_sw_interface (vnm, s->sw_if_index[VLIB_RX]);
907 vlib_increment_combined_counter (im->combined_sw_if_counters
908 + VNET_INTERFACE_COUNTER_RX,
909 vlib_get_thread_index (),
910 si->sw_if_index, n_buffers, length_sum);
916 pg_generate_fix_multi_buffer_lengths (pg_main_t * pg,
918 u32 * buffers, u32 n_buffers)
920 vlib_main_t *vm = vlib_get_main ();
921 pg_buffer_index_t *pbi;
923 static u32 *unused_buffers = 0;
925 while (n_buffers > 0)
931 b = vlib_get_buffer (vm, bi);
933 /* Current length here is length of whole packet. */
934 n_bytes_left = b->current_length;
936 pbi = s->buffer_indices;
939 uword n = clib_min (n_bytes_left, s->buffer_bytes);
941 b->current_length = n;
943 if (n_bytes_left > 0)
944 b->flags |= VLIB_BUFFER_NEXT_PRESENT;
946 b->flags &= ~VLIB_BUFFER_NEXT_PRESENT;
948 /* Return unused buffers to fifos. */
950 vec_add1 (unused_buffers, bi);
953 if (pbi >= vec_end (s->buffer_indices))
957 b = vlib_get_buffer (vm, bi);
959 ASSERT (n_bytes_left == 0);
965 if (vec_len (unused_buffers) > 0)
967 vlib_buffer_free_no_next (vm, unused_buffers, vec_len (unused_buffers));
968 _vec_len (unused_buffers) = 0;
973 pg_generate_edit (pg_main_t * pg,
974 pg_stream_t * s, u32 * buffers, u32 n_buffers)
978 vec_foreach (e, s->non_fixed_edits)
983 case PG_EDIT_INCREMENT:
988 v_min = pg_edit_get_value (e, PG_EDIT_LO);
989 v_max = pg_edit_get_value (e, PG_EDIT_HI);
991 hi_bit = (BITS (u8) * STRUCT_OFFSET_OF (vlib_buffer_t, data)
992 + BITS (u8) + e->lsb_bit_offset);
993 lo_bit = hi_bit - e->n_bits;
995 e->last_increment_value
996 = do_it (pg, s, buffers, n_buffers, lo_bit, hi_bit, v_min, v_max,
997 e->last_increment_value, e->type);
1001 case PG_EDIT_UNSPECIFIED:
1005 /* Should not be any fixed edits left. */
1011 /* Call any edit functions to e.g. completely IP lengths, checksums, ... */
1014 for (i = vec_len (s->edit_groups) - 1; i >= 0; i--)
1016 pg_edit_group_t *g = s->edit_groups + i;
1017 if (g->edit_function)
1018 g->edit_function (pg, s, g, buffers, n_buffers);
1024 pg_set_next_buffer_pointers (pg_main_t * pg,
1026 u32 * buffers, u32 * next_buffers, u32 n_buffers)
1028 vlib_main_t *vm = vlib_get_main ();
1030 while (n_buffers >= 4)
1033 vlib_buffer_t *b0, *b1;
1035 b0 = vlib_get_buffer (vm, buffers[0]);
1036 b1 = vlib_get_buffer (vm, buffers[1]);
1037 ni0 = next_buffers[0];
1038 ni1 = next_buffers[1];
1040 vlib_prefetch_buffer_with_index (vm, buffers[2], WRITE);
1041 vlib_prefetch_buffer_with_index (vm, buffers[3], WRITE);
1043 b0->flags |= VLIB_BUFFER_NEXT_PRESENT;
1044 b1->flags |= VLIB_BUFFER_NEXT_PRESENT;
1045 b0->next_buffer = ni0;
1046 b1->next_buffer = ni1;
1053 while (n_buffers > 0)
1058 b0 = vlib_get_buffer (vm, buffers[0]);
1059 ni0 = next_buffers[0];
1064 b0->flags |= VLIB_BUFFER_NEXT_PRESENT;
1065 b0->next_buffer = ni0;
1069 static_always_inline void
1070 init_buffers_inline (vlib_main_t * vm,
1073 u32 n_buffers, u32 data_offset, u32 n_data, u32 set_data)
1078 ASSERT (s->replay_packet_templates == 0);
1080 data = s->fixed_packet_data + data_offset;
1081 mask = s->fixed_packet_data_mask + data_offset;
1082 if (data + n_data >= vec_end (s->fixed_packet_data))
1083 n_data = (data < vec_end (s->fixed_packet_data)
1084 ? vec_end (s->fixed_packet_data) - data : 0);
1087 ASSERT (data + n_data <= vec_end (s->fixed_packet_data));
1088 ASSERT (mask + n_data <= vec_end (s->fixed_packet_data_mask));
1097 vlib_buffer_t *b0, *b1;
1099 /* Prefetch next iteration. */
1100 vlib_prefetch_buffer_with_index (vm, b[2], STORE);
1101 vlib_prefetch_buffer_with_index (vm, b[3], STORE);
1108 b0 = vlib_get_buffer (vm, bi0);
1109 b1 = vlib_get_buffer (vm, bi1);
1110 b0->flags |= s->buffer_flags;
1111 b1->flags |= s->buffer_flags;
1113 vnet_buffer (b0)->sw_if_index[VLIB_RX] =
1114 vnet_buffer (b1)->sw_if_index[VLIB_RX] = s->sw_if_index[VLIB_RX];
1116 vnet_buffer (b0)->sw_if_index[VLIB_TX] =
1117 vnet_buffer (b1)->sw_if_index[VLIB_TX] = s->sw_if_index[VLIB_TX];
1121 clib_memcpy_fast (b0->data, data, n_data);
1122 clib_memcpy_fast (b1->data, data, n_data);
1126 ASSERT (validate_buffer_data2 (b0, s, data_offset, n_data));
1127 ASSERT (validate_buffer_data2 (b1, s, data_offset, n_data));
1140 b0 = vlib_get_buffer (vm, bi0);
1141 b0->flags |= s->buffer_flags;
1142 vnet_buffer (b0)->sw_if_index[VLIB_RX] = s->sw_if_index[VLIB_RX];
1143 vnet_buffer (b0)->sw_if_index[VLIB_TX] = s->sw_if_index[VLIB_TX];
1146 clib_memcpy_fast (b0->data, data, n_data);
1148 ASSERT (validate_buffer_data2 (b0, s, data_offset, n_data));
1153 pg_stream_fill_helper (pg_main_t * pg,
1155 pg_buffer_index_t * bi,
1156 u32 * buffers, u32 * next_buffers, u32 n_alloc)
1158 vlib_main_t *vm = vlib_get_main ();
1159 uword is_start_of_packet = bi == s->buffer_indices;
1162 ASSERT (vec_len (s->replay_packet_templates) == 0);
1164 n_allocated = vlib_buffer_alloc (vm, buffers, n_alloc);
1165 if (n_allocated == 0)
1169 * We can't assume we got all the buffers we asked for...
1170 * This never worked until recently.
1172 n_alloc = n_allocated;
1174 /* Reinitialize buffers */
1178 n_alloc, (bi - s->buffer_indices) * s->buffer_bytes /* data offset */ ,
1183 pg_set_next_buffer_pointers (pg, s, buffers, next_buffers, n_alloc);
1185 if (is_start_of_packet)
1187 pg_generate_set_lengths (pg, s, buffers, n_alloc);
1188 if (vec_len (s->buffer_indices) > 1)
1189 pg_generate_fix_multi_buffer_lengths (pg, s, buffers, n_alloc);
1191 pg_generate_edit (pg, s, buffers, n_alloc);
1198 pg_stream_fill_replay (pg_main_t * pg, pg_stream_t * s, u32 n_alloc)
1200 pg_buffer_index_t *bi;
1202 u32 buffer_alloc_request = 0;
1203 u32 buffer_alloc_result;
1204 u32 current_buffer_index;
1206 vlib_main_t *vm = vlib_get_main ();
1207 vnet_main_t *vnm = vnet_get_main ();
1208 u32 buf_sz = vlib_buffer_get_default_data_size (vm);
1209 vnet_interface_main_t *im = &vnm->interface_main;
1210 vnet_sw_interface_t *si;
1212 buffers = pg->replay_buffers_by_thread[vm->thread_index];
1213 vec_reset_length (buffers);
1214 bi = s->buffer_indices;
1217 i = s->current_replay_packet_index;
1218 l = vec_len (s->replay_packet_templates);
1220 /* Figure out how many buffers we need */
1225 d0 = vec_elt (s->replay_packet_templates, i);
1226 buffer_alloc_request += (vec_len (d0) + (buf_sz - 1)) / buf_sz;
1228 i = ((i + 1) == l) ? 0 : i + 1;
1232 ASSERT (buffer_alloc_request > 0);
1233 vec_validate (buffers, buffer_alloc_request - 1);
1235 /* Allocate that many buffers */
1236 buffer_alloc_result = vlib_buffer_alloc (vm, buffers, buffer_alloc_request);
1237 if (buffer_alloc_result < buffer_alloc_request)
1239 clib_warning ("alloc failure, got %d not %d", buffer_alloc_result,
1240 buffer_alloc_request);
1241 vlib_buffer_free_no_next (vm, buffers, buffer_alloc_result);
1242 pg->replay_buffers_by_thread[vm->thread_index] = buffers;
1246 /* Now go generate the buffers, and add them to the FIFO */
1249 current_buffer_index = 0;
1250 i = s->current_replay_packet_index;
1251 l = vec_len (s->replay_packet_templates);
1257 u32 bytes_to_copy, bytes_this_chunk;
1260 d0 = vec_elt (s->replay_packet_templates, i);
1262 bytes_to_copy = vec_len (d0);
1264 /* Add head chunk to pg fifo */
1265 clib_fifo_add1 (bi->buffer_fifo, buffers[current_buffer_index]);
1268 while (bytes_to_copy)
1270 bytes_this_chunk = clib_min (bytes_to_copy, buf_sz);
1271 ASSERT (current_buffer_index < vec_len (buffers));
1272 b = vlib_get_buffer (vm, buffers[current_buffer_index]);
1273 clib_memcpy_fast (b->data, d0 + data_offset, bytes_this_chunk);
1274 vnet_buffer (b)->sw_if_index[VLIB_RX] = s->sw_if_index[VLIB_RX];
1275 vnet_buffer (b)->sw_if_index[VLIB_TX] = s->sw_if_index[VLIB_TX];
1276 b->flags = s->buffer_flags;
1278 b->current_data = 0;
1279 b->current_length = bytes_this_chunk;
1281 not_last = bytes_this_chunk < bytes_to_copy;
1284 ASSERT (current_buffer_index < (vec_len (buffers) - 1));
1285 b->flags |= VLIB_BUFFER_NEXT_PRESENT;
1286 b->next_buffer = buffers[current_buffer_index + 1];
1288 bytes_to_copy -= bytes_this_chunk;
1289 data_offset += bytes_this_chunk;
1290 current_buffer_index++;
1293 i = ((i + 1) == l) ? 0 : i + 1;
1297 /* Update the interface counters */
1298 si = vnet_get_sw_interface (vnm, s->sw_if_index[VLIB_RX]);
1300 for (i = 0; i < n_alloc; i++)
1301 l += vlib_buffer_index_length_in_chain (vm, buffers[i]);
1302 vlib_increment_combined_counter (im->combined_sw_if_counters
1303 + VNET_INTERFACE_COUNTER_RX,
1304 vlib_get_thread_index (),
1305 si->sw_if_index, n_alloc, l);
1307 s->current_replay_packet_index += n_alloc;
1308 s->current_replay_packet_index %= vec_len (s->replay_packet_templates);
1310 pg->replay_buffers_by_thread[vm->thread_index] = buffers;
1316 pg_stream_fill (pg_main_t * pg, pg_stream_t * s, u32 n_buffers)
1318 pg_buffer_index_t *bi;
1319 word i, n_in_fifo, n_alloc, n_free, n_added;
1320 u32 *tail, *start, *end, *last_tail, *last_start;
1322 bi = s->buffer_indices;
1324 n_in_fifo = clib_fifo_elts (bi->buffer_fifo);
1325 if (n_in_fifo >= n_buffers)
1328 n_alloc = n_buffers - n_in_fifo;
1330 /* Round up, but never generate more than limit. */
1331 n_alloc = clib_max (VLIB_FRAME_SIZE, n_alloc);
1333 if (s->n_packets_limit > 0
1334 && s->n_packets_generated + n_in_fifo + n_alloc >= s->n_packets_limit)
1336 n_alloc = s->n_packets_limit - s->n_packets_generated - n_in_fifo;
1342 * Handle pcap replay directly
1344 if (s->replay_packet_templates)
1345 return pg_stream_fill_replay (pg, s, n_alloc);
1347 /* All buffer fifos should have the same size. */
1351 vec_foreach (bi, s->buffer_indices)
1353 e = clib_fifo_elts (bi->buffer_fifo);
1354 if (bi == s->buffer_indices)
1360 last_tail = last_start = 0;
1363 for (i = vec_len (s->buffer_indices) - 1; i >= 0; i--)
1365 bi = vec_elt_at_index (s->buffer_indices, i);
1367 n_free = clib_fifo_free_elts (bi->buffer_fifo);
1368 if (n_free < n_alloc)
1369 clib_fifo_resize (bi->buffer_fifo, n_alloc - n_free);
1371 tail = clib_fifo_advance_tail (bi->buffer_fifo, n_alloc);
1372 start = bi->buffer_fifo;
1373 end = clib_fifo_end (bi->buffer_fifo);
1375 if (tail + n_alloc <= end)
1378 pg_stream_fill_helper (pg, s, bi, tail, last_tail, n_alloc);
1382 u32 n = clib_min (end - tail, n_alloc);
1383 n_added = pg_stream_fill_helper (pg, s, bi, tail, last_tail, n);
1385 if (n_added == n && n_alloc > n_added)
1387 n_added += pg_stream_fill_helper
1388 (pg, s, bi, start, last_start, n_alloc - n_added);
1392 if (PREDICT_FALSE (n_added < n_alloc))
1393 tail = clib_fifo_advance_tail (bi->buffer_fifo, n_added - n_alloc);
1398 /* Verify that pkts in the fifo are properly allocated */
1401 return n_in_fifo + n_added;
1411 /* Use pre data for packet data. */
1412 vlib_buffer_t buffer;
1416 format_pg_input_trace (u8 * s, va_list * va)
1418 vlib_main_t *vm = va_arg (*va, vlib_main_t *);
1419 CLIB_UNUSED (vlib_node_t * node) = va_arg (*va, vlib_node_t *);
1420 pg_input_trace_t *t = va_arg (*va, pg_input_trace_t *);
1421 pg_main_t *pg = &pg_main;
1422 pg_stream_t *stream;
1424 u32 indent = format_get_indent (s);
1427 if (!pool_is_free_index (pg->streams, t->stream_index))
1428 stream = pool_elt_at_index (pg->streams, t->stream_index);
1431 s = format (s, "stream %v", pg->streams[t->stream_index].name);
1433 s = format (s, "stream %d", t->stream_index);
1435 s = format (s, ", %d bytes", t->packet_length);
1436 s = format (s, ", sw_if_index %d", t->sw_if_index);
1438 s = format (s, "\n%U%U",
1439 format_white_space, indent, format_vnet_buffer, &t->buffer);
1441 s = format (s, "\n%U", format_white_space, indent);
1445 n = vlib_get_node (vm, stream->node_index);
1447 if (n && n->format_buffer)
1448 s = format (s, "%U", n->format_buffer,
1449 t->buffer.pre_data, sizeof (t->buffer.pre_data));
1451 s = format (s, "%U",
1452 format_hex_bytes, t->buffer.pre_data,
1453 ARRAY_LEN (t->buffer.pre_data));
1458 pg_input_trace (pg_main_t * pg,
1459 vlib_node_runtime_t * node, u32 stream_index, u32 next_index,
1460 u32 * buffers, const u32 n_buffers, const u32 n_trace)
1462 vlib_main_t *vm = vlib_get_main ();
1464 u32 n_trace0 = 0, n_trace1 = 0;
1466 n_left = clib_min (n_buffers, n_trace);
1472 vlib_buffer_t *b0, *b1;
1473 pg_input_trace_t *t0, *t1;
1480 b0 = vlib_get_buffer (vm, bi0);
1481 b1 = vlib_get_buffer (vm, bi1);
1484 vlib_trace_buffer (vm, node, next_index, b0, /* follow_chain */ 1);
1486 vlib_trace_buffer (vm, node, next_index, b1, /* follow_chain */ 1);
1488 t0 = vlib_add_trace (vm, node, b0, sizeof (t0[0]));
1489 t1 = vlib_add_trace (vm, node, b1, sizeof (t1[0]));
1491 t0->stream_index = stream_index;
1492 t1->stream_index = stream_index;
1494 t0->packet_length = vlib_buffer_length_in_chain (vm, b0);
1495 t1->packet_length = vlib_buffer_length_in_chain (vm, b1);
1497 t0->sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX];
1498 t1->sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_RX];
1500 clib_memcpy_fast (&t0->buffer, b0,
1501 sizeof (b0[0]) - sizeof (b0->pre_data));
1502 clib_memcpy_fast (&t1->buffer, b1,
1503 sizeof (b1[0]) - sizeof (b1->pre_data));
1505 clib_memcpy_fast (t0->buffer.pre_data, b0->data,
1506 sizeof (t0->buffer.pre_data));
1507 clib_memcpy_fast (t1->buffer.pre_data, b1->data,
1508 sizeof (t1->buffer.pre_data));
1515 pg_input_trace_t *t0;
1521 b0 = vlib_get_buffer (vm, bi0);
1524 vlib_trace_buffer (vm, node, next_index, b0, /* follow_chain */ 1);
1525 t0 = vlib_add_trace (vm, node, b0, sizeof (t0[0]));
1527 t0->stream_index = stream_index;
1528 t0->packet_length = vlib_buffer_length_in_chain (vm, b0);
1529 t0->sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX];
1530 clib_memcpy_fast (&t0->buffer, b0,
1531 sizeof (b0[0]) - sizeof (b0->pre_data));
1532 clib_memcpy_fast (t0->buffer.pre_data, b0->data,
1533 sizeof (t0->buffer.pre_data));
1536 return n_trace - n_trace0 - n_trace1;
1539 static_always_inline void
1540 fill_buffer_offload_flags (vlib_main_t *vm, u32 *buffers, u32 n_buffers,
1541 u32 buffer_oflags, int gso_enabled, u32 gso_size)
1543 for (int i = 0; i < n_buffers; i++)
1545 vlib_buffer_t *b0 = vlib_get_buffer (vm, buffers[i]);
1547 vnet_buffer_oflags_t oflags = 0;
1549 ethernet_header_t *eh =
1550 (ethernet_header_t *) vlib_buffer_get_current (b0);
1551 u16 ethertype = clib_net_to_host_u16 (eh->type);
1552 u16 l2hdr_sz = sizeof (ethernet_header_t);
1554 if (ethernet_frame_is_tagged (ethertype))
1556 ethernet_vlan_header_t *vlan = (ethernet_vlan_header_t *) (eh + 1);
1558 ethertype = clib_net_to_host_u16 (vlan->type);
1559 l2hdr_sz += sizeof (*vlan);
1560 if (ethertype == ETHERNET_TYPE_VLAN)
1563 ethertype = clib_net_to_host_u16 (vlan->type);
1564 l2hdr_sz += sizeof (*vlan);
1568 vnet_buffer (b0)->l2_hdr_offset = 0;
1569 vnet_buffer (b0)->l3_hdr_offset = l2hdr_sz;
1571 if (PREDICT_TRUE (ethertype == ETHERNET_TYPE_IP4))
1574 (ip4_header_t *) (vlib_buffer_get_current (b0) + l2hdr_sz);
1575 vnet_buffer (b0)->l4_hdr_offset = l2hdr_sz + ip4_header_bytes (ip4);
1576 l4_proto = ip4->protocol;
1578 (VNET_BUFFER_F_IS_IP4 | VNET_BUFFER_F_L2_HDR_OFFSET_VALID |
1579 VNET_BUFFER_F_L3_HDR_OFFSET_VALID |
1580 VNET_BUFFER_F_L4_HDR_OFFSET_VALID);
1581 if (buffer_oflags & VNET_BUFFER_OFFLOAD_F_IP_CKSUM)
1582 oflags |= VNET_BUFFER_OFFLOAD_F_IP_CKSUM;
1584 else if (PREDICT_TRUE (ethertype == ETHERNET_TYPE_IP6))
1587 (ip6_header_t *) (vlib_buffer_get_current (b0) + l2hdr_sz);
1588 vnet_buffer (b0)->l4_hdr_offset = l2hdr_sz + sizeof (ip6_header_t);
1589 /* FIXME IPv6 EH traversal */
1590 l4_proto = ip6->protocol;
1592 (VNET_BUFFER_F_IS_IP6 | VNET_BUFFER_F_L2_HDR_OFFSET_VALID |
1593 VNET_BUFFER_F_L3_HDR_OFFSET_VALID |
1594 VNET_BUFFER_F_L4_HDR_OFFSET_VALID);
1597 if (l4_proto == IP_PROTOCOL_TCP)
1599 if (buffer_oflags & VNET_BUFFER_OFFLOAD_F_TCP_CKSUM)
1600 oflags |= VNET_BUFFER_OFFLOAD_F_TCP_CKSUM;
1602 /* only set GSO flag for chained buffers */
1603 if (gso_enabled && (b0->flags & VLIB_BUFFER_NEXT_PRESENT))
1605 b0->flags |= VNET_BUFFER_F_GSO;
1607 (tcp_header_t *) (vlib_buffer_get_current (b0) +
1608 vnet_buffer (b0)->l4_hdr_offset);
1609 vnet_buffer2 (b0)->gso_l4_hdr_sz = tcp_header_bytes (tcp);
1610 vnet_buffer2 (b0)->gso_size = gso_size;
1613 else if (l4_proto == IP_PROTOCOL_UDP)
1615 if (buffer_oflags & VNET_BUFFER_OFFLOAD_F_UDP_CKSUM)
1616 oflags |= VNET_BUFFER_OFFLOAD_F_UDP_CKSUM;
1620 vnet_buffer_offload_flags_set (b0, oflags);
1625 pg_generate_packets (vlib_node_runtime_t * node,
1627 pg_stream_t * s, uword n_packets_to_generate)
1629 vlib_main_t *vm = vlib_get_main ();
1630 u32 *to_next, n_this_frame, n_left, n_trace, n_packets_in_fifo;
1631 uword n_packets_generated;
1632 pg_buffer_index_t *bi, *bi0;
1633 u32 next_index = s->next_index;
1634 vnet_feature_main_t *fm = &feature_main;
1635 vnet_feature_config_main_t *cm;
1636 u8 feature_arc_index = fm->device_input_feature_arc_index;
1637 cm = &fm->feature_config_mains[feature_arc_index];
1638 u32 current_config_index = ~(u32) 0;
1642 pi = pool_elt_at_index (pg->interfaces,
1643 pg->if_id_by_sw_if_index[s->sw_if_index[VLIB_RX]]);
1644 bi0 = s->buffer_indices;
1646 n_packets_in_fifo = pg_stream_fill (pg, s, n_packets_to_generate);
1647 n_packets_to_generate = clib_min (n_packets_in_fifo, n_packets_to_generate);
1648 n_packets_generated = 0;
1651 (vnet_have_features (feature_arc_index, s->sw_if_index[VLIB_RX])))
1653 current_config_index =
1654 vec_elt (cm->config_index_by_sw_if_index, s->sw_if_index[VLIB_RX]);
1655 vnet_get_config_data (&cm->config_main, ¤t_config_index,
1659 if (PREDICT_FALSE (pi->coalesce_enabled))
1660 vnet_gro_flow_table_schedule_node_on_dispatcher (vm, pi->flow_table);
1662 while (n_packets_to_generate > 0)
1664 u32 *head, *start, *end;
1666 if (PREDICT_TRUE (next_index == VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT))
1668 vlib_next_frame_t *nf;
1670 ethernet_input_frame_t *ef;
1671 vlib_get_new_next_frame (vm, node, next_index, to_next, n_left);
1672 nf = vlib_node_runtime_get_next_frame (vm, node, next_index);
1673 f = vlib_get_frame (vm, nf->frame);
1674 f->flags = ETH_INPUT_FRAME_F_SINGLE_SW_IF_IDX;
1676 ef = vlib_frame_scalar_args (f);
1677 ef->sw_if_index = pi->sw_if_index;
1678 ef->hw_if_index = pi->hw_if_index;
1679 vlib_frame_no_append (f);
1682 vlib_get_next_frame (vm, node, next_index, to_next, n_left);
1684 n_this_frame = n_packets_to_generate;
1685 if (n_this_frame > n_left)
1686 n_this_frame = n_left;
1688 start = bi0->buffer_fifo;
1689 end = clib_fifo_end (bi0->buffer_fifo);
1690 head = clib_fifo_head (bi0->buffer_fifo);
1692 if (head + n_this_frame <= end)
1693 vlib_buffer_copy_indices (to_next, head, n_this_frame);
1697 vlib_buffer_copy_indices (to_next + 0, head, n);
1698 vlib_buffer_copy_indices (to_next + n, start, n_this_frame - n);
1701 if (s->replay_packet_templates == 0)
1703 vec_foreach (bi, s->buffer_indices)
1704 clib_fifo_advance_head (bi->buffer_fifo, n_this_frame);
1708 clib_fifo_advance_head (bi0->buffer_fifo, n_this_frame);
1711 if (current_config_index != ~(u32) 0)
1712 for (i = 0; i < n_this_frame; i++)
1715 b = vlib_get_buffer (vm, to_next[i]);
1716 b->current_config_index = current_config_index;
1717 vnet_buffer (b)->feature_arc_index = feature_arc_index;
1720 if (pi->gso_enabled || (s->buffer_flags & VNET_BUFFER_F_OFFLOAD))
1722 fill_buffer_offload_flags (vm, to_next, n_this_frame,
1723 s->buffer_oflags, pi->gso_enabled,
1727 n_trace = vlib_get_trace_count (vm, node);
1728 if (PREDICT_FALSE (n_trace > 0))
1731 pg_input_trace (pg, node, s - pg->streams, next_index, to_next,
1732 n_this_frame, n_trace);
1733 vlib_set_trace_count (vm, node, n_trace);
1735 n_packets_to_generate -= n_this_frame;
1736 n_packets_generated += n_this_frame;
1737 n_left -= n_this_frame;
1743 for (i = 0; i < n_this_frame; i++)
1745 b = vlib_get_buffer (vm, to_next[i]);
1746 ASSERT ((b->flags & VLIB_BUFFER_NEXT_PRESENT) == 0 ||
1747 b->current_length >= VLIB_BUFFER_MIN_CHAIN_SEG_SIZE);
1750 vlib_put_next_frame (vm, node, next_index, n_left);
1753 return n_packets_generated;
1757 pg_input_stream (vlib_node_runtime_t * node, pg_main_t * pg, pg_stream_t * s)
1759 vlib_main_t *vm = vlib_get_main ();
1763 if (s->n_packets_limit > 0 && s->n_packets_generated >= s->n_packets_limit)
1765 pg_stream_enable_disable (pg, s, /* want_enabled */ 0);
1769 /* Apply rate limit. */
1770 time_now = vlib_time_now (vm);
1771 if (s->time_last_generate == 0)
1772 s->time_last_generate = time_now;
1774 dt = time_now - s->time_last_generate;
1775 s->time_last_generate = time_now;
1777 n_packets = VLIB_FRAME_SIZE;
1778 if (s->rate_packets_per_second > 0)
1780 s->packet_accumulator += dt * s->rate_packets_per_second;
1781 n_packets = s->packet_accumulator;
1783 /* Never allow accumulator to grow if we get behind. */
1784 s->packet_accumulator -= n_packets;
1787 /* Apply fixed limit. */
1788 if (s->n_packets_limit > 0
1789 && s->n_packets_generated + n_packets > s->n_packets_limit)
1790 n_packets = s->n_packets_limit - s->n_packets_generated;
1792 /* Generate up to one frame's worth of packets. */
1793 if (n_packets > s->n_max_frame)
1794 n_packets = s->n_max_frame;
1797 n_packets = pg_generate_packets (node, pg, s, n_packets);
1799 s->n_packets_generated += n_packets;
1805 pg_input (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame)
1808 pg_main_t *pg = &pg_main;
1809 uword n_packets = 0;
1810 u32 worker_index = 0;
1812 if (vlib_num_workers ())
1813 worker_index = vlib_get_current_worker_index ();
1816 clib_bitmap_foreach (i, pg->enabled_streams[worker_index]) {
1817 pg_stream_t *s = vec_elt_at_index (pg->streams, i);
1818 n_packets += pg_input_stream (node, pg, s);
1826 VLIB_REGISTER_NODE (pg_input_node) = {
1827 .function = pg_input,
1828 .flags = VLIB_NODE_FLAG_TRACE_SUPPORTED,
1830 .sibling_of = "device-input",
1831 .type = VLIB_NODE_TYPE_INPUT,
1833 .format_trace = format_pg_input_trace,
1835 /* Input node will be left disabled until a stream is active. */
1836 .state = VLIB_NODE_STATE_DISABLED,
1840 VLIB_NODE_FN (pg_input_mac_filter) (vlib_main_t * vm,
1841 vlib_node_runtime_t * node,
1842 vlib_frame_t * frame)
1844 vlib_buffer_t *bufs[VLIB_FRAME_SIZE], **b = bufs;
1845 u16 nexts[VLIB_FRAME_SIZE], *next;
1846 pg_main_t *pg = &pg_main;
1849 from = vlib_frame_vector_args (frame);
1850 n_left = frame->n_vectors;
1853 clib_memset_u16 (next, 0, VLIB_FRAME_SIZE);
1855 vlib_get_buffers (vm, from, bufs, n_left);
1859 const ethernet_header_t *eth;
1863 pi = pool_elt_at_index
1865 pg->if_id_by_sw_if_index[vnet_buffer (b[0])->sw_if_index[VLIB_RX]]);
1866 eth = vlib_buffer_get_current (b[0]);
1868 mac_address_from_bytes (&in, eth->dst_address);
1870 if (PREDICT_FALSE (ethernet_address_cast (in.bytes)))
1872 mac_address_t *allowed;
1874 if (0 != vec_len (pi->allowed_mcast_macs))
1876 vec_foreach (allowed, pi->allowed_mcast_macs)
1878 if (0 != mac_address_cmp (allowed, &in))
1882 if (vec_is_member (allowed, pi->allowed_mcast_macs))
1883 vnet_feature_next_u16 (&next[0], b[0]);
1892 vlib_buffer_enqueue_to_next (vm, node, from, nexts, frame->n_vectors);
1894 return (frame->n_vectors);
1898 VLIB_REGISTER_NODE (pg_input_mac_filter) = {
1899 .name = "pg-input-mac-filter",
1900 .vector_size = sizeof (u32),
1901 .format_trace = format_pg_input_trace,
1907 VNET_FEATURE_INIT (pg_input_mac_filter_feat, static) = {
1908 .arc_name = "device-input",
1909 .node_name = "pg-input-mac-filter",
1913 static clib_error_t *
1914 pg_input_mac_filter_cfg (vlib_main_t * vm,
1915 unformat_input_t * input, vlib_cli_command_t * cmd)
1917 unformat_input_t _line_input, *line_input = &_line_input;
1918 u32 sw_if_index = ~0;
1921 if (!unformat_user (input, unformat_line_input, line_input))
1924 while (unformat_check_input (line_input) != UNFORMAT_END_OF_INPUT)
1926 if (unformat (line_input, "%U",
1927 unformat_vnet_sw_interface,
1928 vnet_get_main (), &sw_if_index))
1930 else if (unformat (line_input, "%U",
1931 unformat_vlib_enable_disable, &is_enable))
1934 return clib_error_create ("unknown input `%U'",
1935 format_unformat_error, line_input);
1937 unformat_free (line_input);
1939 if (~0 == sw_if_index)
1940 return clib_error_create ("specify interface");
1942 vnet_feature_enable_disable ("device-input",
1943 "pg-input-mac-filter",
1944 sw_if_index, is_enable, 0, 0);
1950 VLIB_CLI_COMMAND (enable_streams_cli, static) = {
1951 .path = "packet-generator mac-filter",
1952 .short_help = "packet-generator mac-filter <INTERFACE> <on|off>",
1953 .function = pg_input_mac_filter_cfg,
1959 * fd.io coding-style-patch-verification: ON
1962 * eval: (c-set-style "gnu")