2 * Copyright (c) 2015 Cisco and/or its affiliates.
3 * Licensed under the Apache License, Version 2.0 (the "License");
4 * you may not use this file except in compliance with the License.
5 * You may obtain a copy of the License at:
7 * http://www.apache.org/licenses/LICENSE-2.0
9 * Unless required by applicable law or agreed to in writing, software
10 * distributed under the License is distributed on an "AS IS" BASIS,
11 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 * See the License for the specific language governing permissions and
13 * limitations under the License.
16 * pg_input.c: buffer generator input
18 * Copyright (c) 2008 Eliot Dresselhaus
20 * Permission is hereby granted, free of charge, to any person obtaining
21 * a copy of this software and associated documentation files (the
22 * "Software"), to deal in the Software without restriction, including
23 * without limitation the rights to use, copy, modify, merge, publish,
24 * distribute, sublicense, and/or sell copies of the Software, and to
25 * permit persons to whom the Software is furnished to do so, subject to
26 * the following conditions:
28 * The above copyright notice and this permission notice shall be
29 * included in all copies or substantial portions of the Software.
31 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
32 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
33 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
34 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
35 * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
36 * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
37 * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
40 #include <vlib/vlib.h>
41 #include <vnet/pg/pg.h>
42 #include <vnet/vnet.h>
43 #include <vnet/feature/feature.h>
44 #include <vnet/devices/devices.h>
47 validate_buffer_data2 (vlib_buffer_t * b, pg_stream_t * s,
48 u32 data_offset, u32 n_bytes)
54 pd = s->fixed_packet_data + data_offset;
55 pm = s->fixed_packet_data_mask + data_offset;
57 if (pd + n_bytes >= vec_end (s->fixed_packet_data))
58 n_bytes = (pd < vec_end (s->fixed_packet_data)
59 ? vec_end (s->fixed_packet_data) - pd : 0);
61 for (i = 0; i < n_bytes; i++)
62 if ((bd[i] & pm[i]) != pd[i])
68 clib_warning ("buffer %U", format_vlib_buffer, b);
69 clib_warning ("differ at index %d", i);
70 clib_warning ("is %U", format_hex_bytes, bd, n_bytes);
71 clib_warning ("mask %U", format_hex_bytes, pm, n_bytes);
72 clib_warning ("expect %U", format_hex_bytes, pd, n_bytes);
77 validate_buffer_data (vlib_buffer_t * b, pg_stream_t * s)
79 return validate_buffer_data2 (b, s, 0, s->buffer_bytes);
84 u64 v0, u64 v_min, u64 v_max, u32 n_bits, u32 is_net_byte_order)
86 ASSERT (v0 >= v_min && v0 <= v_max);
87 if (n_bits == BITS (u8))
91 else if (n_bits == BITS (u16))
93 if (is_net_byte_order)
94 v0 = clib_host_to_net_u16 (v0);
95 clib_mem_unaligned (a0, u16) = v0;
97 else if (n_bits == BITS (u32))
99 if (is_net_byte_order)
100 v0 = clib_host_to_net_u32 (v0);
101 clib_mem_unaligned (a0, u32) = v0;
103 else if (n_bits == BITS (u64))
105 if (is_net_byte_order)
106 v0 = clib_host_to_net_u64 (v0);
107 clib_mem_unaligned (a0, u64) = v0;
112 set_2 (void *a0, void *a1,
114 u64 v_min, u64 v_max,
115 u32 n_bits, u32 is_net_byte_order, u32 is_increment)
117 ASSERT (v0 >= v_min && v0 <= v_max);
118 ASSERT (v1 >= v_min && v1 <= (v_max + is_increment));
119 if (n_bits == BITS (u8))
124 else if (n_bits == BITS (u16))
126 if (is_net_byte_order)
128 v0 = clib_host_to_net_u16 (v0);
129 v1 = clib_host_to_net_u16 (v1);
131 clib_mem_unaligned (a0, u16) = v0;
132 clib_mem_unaligned (a1, u16) = v1;
134 else if (n_bits == BITS (u32))
136 if (is_net_byte_order)
138 v0 = clib_host_to_net_u32 (v0);
139 v1 = clib_host_to_net_u32 (v1);
141 clib_mem_unaligned (a0, u32) = v0;
142 clib_mem_unaligned (a1, u32) = v1;
144 else if (n_bits == BITS (u64))
146 if (is_net_byte_order)
148 v0 = clib_host_to_net_u64 (v0);
149 v1 = clib_host_to_net_u64 (v1);
151 clib_mem_unaligned (a0, u64) = v0;
152 clib_mem_unaligned (a1, u64) = v1;
156 static_always_inline void
157 do_set_fixed (pg_main_t * pg,
162 u32 byte_offset, u32 is_net_byte_order, u64 v_min, u64 v_max)
164 vlib_main_t *vm = vlib_get_main ();
166 while (n_buffers >= 4)
168 vlib_buffer_t *b0, *b1, *b2, *b3;
171 b0 = vlib_get_buffer (vm, buffers[0]);
172 b1 = vlib_get_buffer (vm, buffers[1]);
173 b2 = vlib_get_buffer (vm, buffers[2]);
174 b3 = vlib_get_buffer (vm, buffers[3]);
178 a0 = (void *) b0 + byte_offset;
179 a1 = (void *) b1 + byte_offset;
180 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
181 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
183 set_2 (a0, a1, v_min, v_min, v_min, v_max, n_bits, is_net_byte_order,
184 /* is_increment */ 0);
186 ASSERT (validate_buffer_data (b0, s));
187 ASSERT (validate_buffer_data (b1, s));
190 while (n_buffers > 0)
195 b0 = vlib_get_buffer (vm, buffers[0]);
199 a0 = (void *) b0 + byte_offset;
201 set_1 (a0, v_min, v_min, v_max, n_bits, is_net_byte_order);
203 ASSERT (validate_buffer_data (b0, s));
207 static_always_inline u64
208 do_set_increment (pg_main_t * pg,
214 u32 is_net_byte_order,
215 u32 want_sum, u64 * sum_result, u64 v_min, u64 v_max, u64 v)
217 vlib_main_t *vm = vlib_get_main ();
220 ASSERT (v >= v_min && v <= v_max);
222 while (n_buffers >= 4)
224 vlib_buffer_t *b0, *b1, *b2, *b3;
228 b0 = vlib_get_buffer (vm, buffers[0]);
229 b1 = vlib_get_buffer (vm, buffers[1]);
230 b2 = vlib_get_buffer (vm, buffers[2]);
231 b3 = vlib_get_buffer (vm, buffers[3]);
235 a0 = (void *) b0 + byte_offset;
236 a1 = (void *) b1 + byte_offset;
237 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
238 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
242 v = v > v_max ? v_min : v;
244 v_old + 0, v_old + 1, v_min, v_max, n_bits, is_net_byte_order,
245 /* is_increment */ 1);
248 sum += 2 * v_old + 1;
250 if (PREDICT_FALSE (v_old + 1 > v_max))
253 sum -= 2 * v_old + 1;
256 set_1 (a0, v + 0, v_min, v_max, n_bits, is_net_byte_order);
261 v = v > v_max ? v_min : v;
262 set_1 (a1, v + 0, v_min, v_max, n_bits, is_net_byte_order);
268 ASSERT (validate_buffer_data (b0, s));
269 ASSERT (validate_buffer_data (b1, s));
272 while (n_buffers > 0)
278 b0 = vlib_get_buffer (vm, buffers[0]);
282 a0 = (void *) b0 + byte_offset;
288 v = v > v_max ? v_min : v;
290 ASSERT (v_old >= v_min && v_old <= v_max);
291 set_1 (a0, v_old, v_min, v_max, n_bits, is_net_byte_order);
293 ASSERT (validate_buffer_data (b0, s));
302 static_always_inline void
303 do_set_random (pg_main_t * pg,
309 u32 is_net_byte_order,
310 u32 want_sum, u64 * sum_result, u64 v_min, u64 v_max)
312 vlib_main_t *vm = vlib_get_main ();
313 u64 v_diff = v_max - v_min + 1;
314 u64 r_mask = max_pow2 (v_diff) - 1;
319 random_data = clib_random_buffer_get_data
320 (&vm->random_buffer, n_buffers * n_bits / BITS (u8));
324 while (n_buffers >= 4)
326 vlib_buffer_t *b0, *b1, *b2, *b3;
328 u64 r0 = 0, r1 = 0; /* warnings be gone */
330 b0 = vlib_get_buffer (vm, buffers[0]);
331 b1 = vlib_get_buffer (vm, buffers[1]);
332 b2 = vlib_get_buffer (vm, buffers[2]);
333 b3 = vlib_get_buffer (vm, buffers[3]);
337 a0 = (void *) b0 + byte_offset;
338 a1 = (void *) b1 + byte_offset;
339 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
340 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
347 u##n * r = random_data; \
350 random_data = r + 2; \
362 /* Add power of 2 sized random number which may be out of range. */
366 /* Twice should be enough to reduce to v_min .. v_max range. */
367 v0 = v0 > v_max ? v0 - v_diff : v0;
368 v1 = v1 > v_max ? v1 - v_diff : v1;
369 v0 = v0 > v_max ? v0 - v_diff : v0;
370 v1 = v1 > v_max ? v1 - v_diff : v1;
375 set_2 (a0, a1, v0, v1, v_min, v_max, n_bits, is_net_byte_order,
376 /* is_increment */ 0);
378 ASSERT (validate_buffer_data (b0, s));
379 ASSERT (validate_buffer_data (b1, s));
382 while (n_buffers > 0)
386 u64 r0 = 0; /* warnings be gone */
388 b0 = vlib_get_buffer (vm, buffers[0]);
392 a0 = (void *) b0 + byte_offset;
399 u##n * r = random_data; \
401 random_data = r + 1; \
413 /* Add power of 2 sized random number which may be out of range. */
416 /* Twice should be enough to reduce to v_min .. v_max range. */
417 v0 = v0 > v_max ? v0 - v_diff : v0;
418 v0 = v0 > v_max ? v0 - v_diff : v0;
423 set_1 (a0, v0, v_min, v_max, n_bits, is_net_byte_order);
425 ASSERT (validate_buffer_data (b0, s));
433 clib_mem_unaligned (a##i, t) = \
434 clib_host_to_net_##t ((clib_net_to_host_mem_##t (a##i) &~ mask) \
440 u64 v_min, u64 v_max,
441 u32 max_bits, u32 n_bits, u64 mask, u32 shift)
443 ASSERT (v0 >= v_min && v0 <= v_max);
444 if (max_bits == BITS (u8))
445 ((u8 *) a0)[0] = (((u8 *) a0)[0] & ~mask) | (v0 << shift);
447 else if (max_bits == BITS (u16))
451 else if (max_bits == BITS (u32))
455 else if (max_bits == BITS (u64))
462 setbits_2 (void *a0, void *a1,
464 u64 v_min, u64 v_max,
465 u32 max_bits, u32 n_bits, u64 mask, u32 shift, u32 is_increment)
467 ASSERT (v0 >= v_min && v0 <= v_max);
468 ASSERT (v1 >= v_min && v1 <= v_max + is_increment);
469 if (max_bits == BITS (u8))
471 ((u8 *) a0)[0] = (((u8 *) a0)[0] & ~mask) | (v0 << shift);
472 ((u8 *) a1)[0] = (((u8 *) a1)[0] & ~mask) | (v1 << shift);
475 else if (max_bits == BITS (u16))
480 else if (max_bits == BITS (u32))
485 else if (max_bits == BITS (u64))
494 static_always_inline void
495 do_setbits_fixed (pg_main_t * pg,
501 u32 byte_offset, u64 v_min, u64 v_max, u64 mask, u32 shift)
503 vlib_main_t *vm = vlib_get_main ();
505 while (n_buffers >= 4)
507 vlib_buffer_t *b0, *b1, *b2, *b3;
510 b0 = vlib_get_buffer (vm, buffers[0]);
511 b1 = vlib_get_buffer (vm, buffers[1]);
512 b2 = vlib_get_buffer (vm, buffers[2]);
513 b3 = vlib_get_buffer (vm, buffers[3]);
517 a0 = (void *) b0 + byte_offset;
518 a1 = (void *) b1 + byte_offset;
519 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
520 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
523 v_min, v_min, v_min, v_max, max_bits, n_bits, mask, shift,
524 /* is_increment */ 0);
526 ASSERT (validate_buffer_data (b0, s));
527 ASSERT (validate_buffer_data (b1, s));
530 while (n_buffers > 0)
535 b0 = vlib_get_buffer (vm, buffers[0]);
539 a0 = (void *) b0 + byte_offset;
541 setbits_1 (a0, v_min, v_min, v_max, max_bits, n_bits, mask, shift);
542 ASSERT (validate_buffer_data (b0, s));
546 static_always_inline u64
547 do_setbits_increment (pg_main_t * pg,
554 u64 v_min, u64 v_max, u64 v, u64 mask, u32 shift)
556 vlib_main_t *vm = vlib_get_main ();
558 ASSERT (v >= v_min && v <= v_max);
560 while (n_buffers >= 4)
562 vlib_buffer_t *b0, *b1, *b2, *b3;
566 b0 = vlib_get_buffer (vm, buffers[0]);
567 b1 = vlib_get_buffer (vm, buffers[1]);
568 b2 = vlib_get_buffer (vm, buffers[2]);
569 b3 = vlib_get_buffer (vm, buffers[3]);
573 a0 = (void *) b0 + byte_offset;
574 a1 = (void *) b1 + byte_offset;
575 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
576 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
580 v = v > v_max ? v_min : v;
582 v_old + 0, v_old + 1,
583 v_min, v_max, max_bits, n_bits, mask, shift,
584 /* is_increment */ 1);
586 if (PREDICT_FALSE (v_old + 1 > v_max))
589 setbits_1 (a0, v + 0, v_min, v_max, max_bits, n_bits, mask, shift);
592 v = v > v_max ? v_min : v;
593 setbits_1 (a1, v + 0, v_min, v_max, max_bits, n_bits, mask, shift);
596 ASSERT (validate_buffer_data (b0, s));
597 ASSERT (validate_buffer_data (b1, s));
600 while (n_buffers > 0)
606 b0 = vlib_get_buffer (vm, buffers[0]);
610 a0 = (void *) b0 + byte_offset;
614 v = v > v_max ? v_min : v;
616 ASSERT (v_old >= v_min && v_old <= v_max);
617 setbits_1 (a0, v_old, v_min, v_max, max_bits, n_bits, mask, shift);
619 ASSERT (validate_buffer_data (b0, s));
625 static_always_inline void
626 do_setbits_random (pg_main_t * pg,
632 u32 byte_offset, u64 v_min, u64 v_max, u64 mask, u32 shift)
634 vlib_main_t *vm = vlib_get_main ();
635 u64 v_diff = v_max - v_min + 1;
636 u64 r_mask = max_pow2 (v_diff) - 1;
640 random_data = clib_random_buffer_get_data
641 (&vm->random_buffer, n_buffers * max_bits / BITS (u8));
644 while (n_buffers >= 4)
646 vlib_buffer_t *b0, *b1, *b2, *b3;
648 u64 r0 = 0, r1 = 0; /* warnings be gone */
650 b0 = vlib_get_buffer (vm, buffers[0]);
651 b1 = vlib_get_buffer (vm, buffers[1]);
652 b2 = vlib_get_buffer (vm, buffers[2]);
653 b3 = vlib_get_buffer (vm, buffers[3]);
657 a0 = (void *) b0 + byte_offset;
658 a1 = (void *) b1 + byte_offset;
659 CLIB_PREFETCH ((void *) b2 + byte_offset, sizeof (v_min), WRITE);
660 CLIB_PREFETCH ((void *) b3 + byte_offset, sizeof (v_min), WRITE);
667 u##n * r = random_data; \
670 random_data = r + 2; \
682 /* Add power of 2 sized random number which may be out of range. */
686 /* Twice should be enough to reduce to v_min .. v_max range. */
687 v0 = v0 > v_max ? v0 - v_diff : v0;
688 v1 = v1 > v_max ? v1 - v_diff : v1;
689 v0 = v0 > v_max ? v0 - v_diff : v0;
690 v1 = v1 > v_max ? v1 - v_diff : v1;
692 setbits_2 (a0, a1, v0, v1, v_min, v_max, max_bits, n_bits, mask, shift,
693 /* is_increment */ 0);
695 ASSERT (validate_buffer_data (b0, s));
696 ASSERT (validate_buffer_data (b1, s));
699 while (n_buffers > 0)
703 u64 r0 = 0; /* warnings be gone */
705 b0 = vlib_get_buffer (vm, buffers[0]);
709 a0 = (void *) b0 + byte_offset;
716 u##n * r = random_data; \
718 random_data = r + 1; \
730 /* Add power of 2 sized random number which may be out of range. */
733 /* Twice should be enough to reduce to v_min .. v_max range. */
734 v0 = v0 > v_max ? v0 - v_diff : v0;
735 v0 = v0 > v_max ? v0 - v_diff : v0;
737 setbits_1 (a0, v0, v_min, v_max, max_bits, n_bits, mask, shift);
739 ASSERT (validate_buffer_data (b0, s));
744 do_it (pg_main_t * pg,
748 u32 lo_bit, u32 hi_bit,
749 u64 v_min, u64 v_max, u64 v, pg_edit_type_t edit_type)
751 u32 max_bits, l0, l1, h1, start_bit;
754 edit_type = PG_EDIT_FIXED;
756 l0 = lo_bit / BITS (u8);
757 l1 = lo_bit % BITS (u8);
758 h1 = hi_bit % BITS (u8);
760 start_bit = l0 * BITS (u8);
762 max_bits = hi_bit - start_bit;
763 ASSERT (max_bits <= 64);
767 if (edit_type == PG_EDIT_INCREMENT) \
768 v = do_set_increment (pg, s, buffers, n_buffers, \
771 /* is_net_byte_order */ 1, \
772 /* want sum */ 0, 0, \
775 else if (edit_type == PG_EDIT_RANDOM) \
776 do_set_random (pg, s, buffers, n_buffers, \
779 /* is_net_byte_order */ 1, \
780 /* want sum */ 0, 0, \
782 else /* edit_type == PG_EDIT_FIXED */ \
783 do_set_fixed (pg, s, buffers, n_buffers, \
786 /* is_net_byte_order */ 1, \
790 if (l1 == 0 && h1 == 0)
806 u32 n_bits = max_bits;
808 max_bits = clib_max (max_pow2 (n_bits), 8);
810 mask = ((u64) 1 << (u64) n_bits) - 1;
811 mask &= ~(((u64) 1 << (u64) shift) - 1);
813 mask <<= max_bits - n_bits;
814 shift += max_bits - n_bits;
820 if (edit_type == PG_EDIT_INCREMENT) \
821 v = do_setbits_increment (pg, s, buffers, n_buffers, \
822 BITS (u##n), n_bits, \
823 l0, v_min, v_max, v, \
825 else if (edit_type == PG_EDIT_RANDOM) \
826 do_setbits_random (pg, s, buffers, n_buffers, \
827 BITS (u##n), n_bits, \
830 else /* edit_type == PG_EDIT_FIXED */ \
831 do_setbits_fixed (pg, s, buffers, n_buffers, \
832 BITS (u##n), n_bits, \
851 pg_generate_set_lengths (pg_main_t * pg,
852 pg_stream_t * s, u32 * buffers, u32 n_buffers)
854 u64 v_min, v_max, length_sum;
855 pg_edit_type_t edit_type;
857 v_min = s->min_packet_bytes;
858 v_max = s->max_packet_bytes;
859 edit_type = s->packet_size_edit_type;
861 if (edit_type == PG_EDIT_INCREMENT)
862 s->last_increment_packet_size
863 = do_set_increment (pg, s, buffers, n_buffers,
864 8 * STRUCT_SIZE_OF (vlib_buffer_t, current_length),
865 STRUCT_OFFSET_OF (vlib_buffer_t, current_length),
866 /* is_net_byte_order */ 0,
867 /* want sum */ 1, &length_sum,
868 v_min, v_max, s->last_increment_packet_size);
870 else if (edit_type == PG_EDIT_RANDOM)
871 do_set_random (pg, s, buffers, n_buffers,
872 8 * STRUCT_SIZE_OF (vlib_buffer_t, current_length),
873 STRUCT_OFFSET_OF (vlib_buffer_t, current_length),
874 /* is_net_byte_order */ 0,
875 /* want sum */ 1, &length_sum,
878 else /* edit_type == PG_EDIT_FIXED */
880 do_set_fixed (pg, s, buffers, n_buffers,
881 8 * STRUCT_SIZE_OF (vlib_buffer_t, current_length),
882 STRUCT_OFFSET_OF (vlib_buffer_t, current_length),
883 /* is_net_byte_order */ 0,
885 length_sum = v_min * n_buffers;
889 vnet_main_t *vnm = vnet_get_main ();
890 vnet_interface_main_t *im = &vnm->interface_main;
891 vnet_sw_interface_t *si =
892 vnet_get_sw_interface (vnm, s->sw_if_index[VLIB_RX]);
894 vlib_increment_combined_counter (im->combined_sw_if_counters
895 + VNET_INTERFACE_COUNTER_RX,
896 os_get_cpu_number (),
897 si->sw_if_index, n_buffers, length_sum);
903 pg_generate_fix_multi_buffer_lengths (pg_main_t * pg,
905 u32 * buffers, u32 n_buffers)
907 vlib_main_t *vm = vlib_get_main ();
908 pg_buffer_index_t *pbi;
910 static u32 *unused_buffers = 0;
912 while (n_buffers > 0)
918 b = vlib_get_buffer (vm, bi);
920 /* Current length here is length of whole packet. */
921 n_bytes_left = b->current_length;
923 pbi = s->buffer_indices;
926 uword n = clib_min (n_bytes_left, s->buffer_bytes);
928 b->current_length = n;
930 if (n_bytes_left > 0)
931 b->flags |= VLIB_BUFFER_NEXT_PRESENT;
933 b->flags &= ~VLIB_BUFFER_NEXT_PRESENT;
935 /* Return unused buffers to fifos. */
937 vec_add1 (unused_buffers, bi);
940 if (pbi >= vec_end (s->buffer_indices))
944 b = vlib_get_buffer (vm, bi);
946 ASSERT (n_bytes_left == 0);
952 if (vec_len (unused_buffers) > 0)
954 vlib_buffer_free_no_next (vm, unused_buffers, vec_len (unused_buffers));
955 _vec_len (unused_buffers) = 0;
960 pg_generate_edit (pg_main_t * pg,
961 pg_stream_t * s, u32 * buffers, u32 n_buffers)
965 vec_foreach (e, s->non_fixed_edits)
970 case PG_EDIT_INCREMENT:
975 v_min = pg_edit_get_value (e, PG_EDIT_LO);
976 v_max = pg_edit_get_value (e, PG_EDIT_HI);
978 hi_bit = (BITS (u8) * STRUCT_OFFSET_OF (vlib_buffer_t, data)
979 + BITS (u8) + e->lsb_bit_offset);
980 lo_bit = hi_bit - e->n_bits;
982 e->last_increment_value
983 = do_it (pg, s, buffers, n_buffers, lo_bit, hi_bit, v_min, v_max,
984 e->last_increment_value, e->type);
988 case PG_EDIT_UNSPECIFIED:
992 /* Should not be any fixed edits left. */
998 /* Call any edit functions to e.g. completely IP lengths, checksums, ... */
1001 for (i = vec_len (s->edit_groups) - 1; i >= 0; i--)
1003 pg_edit_group_t *g = s->edit_groups + i;
1004 if (g->edit_function)
1005 g->edit_function (pg, s, g, buffers, n_buffers);
1011 pg_set_next_buffer_pointers (pg_main_t * pg,
1013 u32 * buffers, u32 * next_buffers, u32 n_buffers)
1015 vlib_main_t *vm = vlib_get_main ();
1017 while (n_buffers >= 4)
1020 vlib_buffer_t *b0, *b1;
1022 b0 = vlib_get_buffer (vm, buffers[0]);
1023 b1 = vlib_get_buffer (vm, buffers[1]);
1024 ni0 = next_buffers[0];
1025 ni1 = next_buffers[1];
1027 vlib_prefetch_buffer_with_index (vm, buffers[2], WRITE);
1028 vlib_prefetch_buffer_with_index (vm, buffers[3], WRITE);
1030 b0->flags |= VLIB_BUFFER_NEXT_PRESENT;
1031 b1->flags |= VLIB_BUFFER_NEXT_PRESENT;
1032 b0->next_buffer = ni0;
1033 b1->next_buffer = ni1;
1040 while (n_buffers > 0)
1045 b0 = vlib_get_buffer (vm, buffers[0]);
1046 ni0 = next_buffers[0];
1051 b0->flags |= VLIB_BUFFER_NEXT_PRESENT;
1052 b0->next_buffer = ni0;
1056 static_always_inline void
1057 init_replay_buffers_inline (vlib_main_t * vm,
1060 u32 n_buffers, u32 data_offset, u32 n_data)
1062 u32 n_left, *b, i, l;
1066 i = s->current_replay_packet_index;
1067 l = vec_len (s->replay_packet_templates);
1079 b0 = vlib_get_buffer (vm, bi0);
1081 vnet_buffer (b0)->sw_if_index[VLIB_RX] = s->sw_if_index[VLIB_RX];
1082 /* was s->sw_if_index[VLIB_TX]; */
1083 vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
1085 d0 = vec_elt (s->replay_packet_templates, i);
1088 if (data_offset + n_data >= vec_len (d0))
1089 n0 = vec_len (d0) > data_offset ? vec_len (d0) - data_offset : 0;
1091 b0->current_length = n0;
1093 clib_memcpy (b0->data, d0 + data_offset, n0);
1094 i = i + 1 == l ? 0 : i + 1;
1098 static_always_inline void
1099 init_buffers_inline (vlib_main_t * vm,
1102 u32 n_buffers, u32 data_offset, u32 n_data, u32 set_data)
1107 if (vec_len (s->replay_packet_templates) > 0)
1108 return init_replay_buffers_inline (vm, s, buffers, n_buffers, data_offset,
1111 data = s->fixed_packet_data + data_offset;
1112 mask = s->fixed_packet_data_mask + data_offset;
1113 if (data + n_data >= vec_end (s->fixed_packet_data))
1114 n_data = (data < vec_end (s->fixed_packet_data)
1115 ? vec_end (s->fixed_packet_data) - data : 0);
1118 ASSERT (data + n_data <= vec_end (s->fixed_packet_data));
1119 ASSERT (mask + n_data <= vec_end (s->fixed_packet_data_mask));
1128 vlib_buffer_t *b0, *b1;
1130 /* Prefetch next iteration. */
1131 vlib_prefetch_buffer_with_index (vm, b[2], STORE);
1132 vlib_prefetch_buffer_with_index (vm, b[3], STORE);
1139 b0 = vlib_get_buffer (vm, bi0);
1140 b1 = vlib_get_buffer (vm, bi1);
1142 vnet_buffer (b0)->sw_if_index[VLIB_RX] =
1143 vnet_buffer (b1)->sw_if_index[VLIB_RX] = s->sw_if_index[VLIB_RX];
1145 vnet_buffer (b0)->sw_if_index[VLIB_TX] =
1146 vnet_buffer (b1)->sw_if_index[VLIB_TX] = (u32) ~ 0;
1150 clib_memcpy (b0->data, data, n_data);
1151 clib_memcpy (b1->data, data, n_data);
1155 ASSERT (validate_buffer_data2 (b0, s, data_offset, n_data));
1156 ASSERT (validate_buffer_data2 (b1, s, data_offset, n_data));
1169 b0 = vlib_get_buffer (vm, bi0);
1170 vnet_buffer (b0)->sw_if_index[VLIB_RX] = s->sw_if_index[VLIB_RX];
1171 /* s->sw_if_index[VLIB_TX]; */
1172 vnet_buffer (b0)->sw_if_index[VLIB_TX] = (u32) ~ 0;
1175 clib_memcpy (b0->data, data, n_data);
1177 ASSERT (validate_buffer_data2 (b0, s, data_offset, n_data));
1182 pg_buffer_init (vlib_main_t * vm,
1183 vlib_buffer_free_list_t * fl, u32 * buffers, u32 n_buffers)
1185 pg_main_t *pg = &pg_main;
1189 si = fl->buffer_init_function_opaque & pow2_mask (24);
1190 bi = fl->buffer_init_function_opaque >> 24;
1192 s = pool_elt_at_index (pg->streams, si);
1194 init_buffers_inline (vm, s, buffers, n_buffers,
1195 /* data_offset */ bi * s->buffer_bytes,
1196 /* n_data */ s->buffer_bytes,
1201 pg_stream_fill_helper (pg_main_t * pg,
1203 pg_buffer_index_t * bi,
1204 u32 * buffers, u32 * next_buffers, u32 n_alloc)
1206 vlib_main_t *vm = vlib_get_main ();
1207 vlib_buffer_free_list_t *f;
1208 uword is_start_of_packet = bi == s->buffer_indices;
1211 f = vlib_buffer_get_free_list (vm, bi->free_list_index);
1214 * Historically, the pg maintained its own free lists and
1215 * device drivers tx paths would return pkts.
1217 if (vm->buffer_main->extern_buffer_mgmt == 0 &&
1218 !(s->flags & PG_STREAM_FLAGS_DISABLE_BUFFER_RECYCLE))
1219 f->buffer_init_function = pg_buffer_init;
1220 f->buffer_init_function_opaque =
1221 (s - pg->streams) | ((bi - s->buffer_indices) << 24);
1223 if (is_start_of_packet)
1224 vnet_buffer (&f->buffer_init_template)->sw_if_index[VLIB_RX]
1225 = vnet_main.local_interface_sw_if_index;
1227 n_allocated = vlib_buffer_alloc_from_free_list (vm,
1230 bi->free_list_index);
1231 if (n_allocated == 0)
1235 * We can't assume we got all the buffers we asked for...
1236 * This never worked until recently.
1238 n_alloc = n_allocated;
1240 /* Reinitialize buffers */
1241 if (vm->buffer_main->extern_buffer_mgmt == 0 || CLIB_DEBUG > 0
1242 || (s->flags & PG_STREAM_FLAGS_DISABLE_BUFFER_RECYCLE))
1246 n_alloc, (bi - s->buffer_indices) * s->buffer_bytes /* data offset */ ,
1249 vm->buffer_main->extern_buffer_mgmt != 0
1250 || (s->flags & PG_STREAM_FLAGS_DISABLE_BUFFER_RECYCLE) != 0);
1253 pg_set_next_buffer_pointers (pg, s, buffers, next_buffers, n_alloc);
1255 if (is_start_of_packet)
1257 if (vec_len (s->replay_packet_templates) > 0)
1259 vnet_main_t *vnm = vnet_get_main ();
1260 vnet_interface_main_t *im = &vnm->interface_main;
1261 vnet_sw_interface_t *si =
1262 vnet_get_sw_interface (vnm, s->sw_if_index[VLIB_RX]);
1265 for (i = 0; i < n_alloc; i++)
1266 l += vlib_buffer_index_length_in_chain (vm, buffers[i]);
1267 vlib_increment_combined_counter (im->combined_sw_if_counters
1268 + VNET_INTERFACE_COUNTER_RX,
1269 os_get_cpu_number (),
1270 si->sw_if_index, n_alloc, l);
1271 s->current_replay_packet_index += n_alloc;
1272 s->current_replay_packet_index %=
1273 vec_len (s->replay_packet_templates);
1277 pg_generate_set_lengths (pg, s, buffers, n_alloc);
1278 if (vec_len (s->buffer_indices) > 1)
1279 pg_generate_fix_multi_buffer_lengths (pg, s, buffers, n_alloc);
1281 pg_generate_edit (pg, s, buffers, n_alloc);
1289 pg_stream_fill (pg_main_t * pg, pg_stream_t * s, u32 n_buffers)
1291 pg_buffer_index_t *bi;
1292 word i, n_in_fifo, n_alloc, n_free, n_added;
1293 u32 *tail, *start, *end, *last_tail, *last_start;
1295 bi = s->buffer_indices;
1297 n_in_fifo = clib_fifo_elts (bi->buffer_fifo);
1298 if (n_in_fifo >= n_buffers)
1301 n_alloc = n_buffers - n_in_fifo;
1303 /* Round up, but never generate more than limit. */
1304 n_alloc = clib_max (VLIB_FRAME_SIZE, n_alloc);
1306 if (s->n_packets_limit > 0
1307 && s->n_packets_generated + n_in_fifo + n_alloc >= s->n_packets_limit)
1309 n_alloc = s->n_packets_limit - s->n_packets_generated - n_in_fifo;
1314 /* All buffer fifos should have the same size. */
1318 vec_foreach (bi, s->buffer_indices)
1320 e = clib_fifo_elts (bi->buffer_fifo);
1321 if (bi == s->buffer_indices)
1327 last_tail = last_start = 0;
1330 for (i = vec_len (s->buffer_indices) - 1; i >= 0; i--)
1332 bi = vec_elt_at_index (s->buffer_indices, i);
1334 n_free = clib_fifo_free_elts (bi->buffer_fifo);
1335 if (n_free < n_alloc)
1336 clib_fifo_resize (bi->buffer_fifo, n_alloc - n_free);
1338 tail = clib_fifo_advance_tail (bi->buffer_fifo, n_alloc);
1339 start = bi->buffer_fifo;
1340 end = clib_fifo_end (bi->buffer_fifo);
1342 if (tail + n_alloc <= end)
1345 pg_stream_fill_helper (pg, s, bi, tail, last_tail, n_alloc);
1349 u32 n = clib_min (end - tail, n_alloc);
1350 n_added = pg_stream_fill_helper (pg, s, bi, tail, last_tail, n);
1352 if (n_added == n && n_alloc > n_added)
1354 n_added += pg_stream_fill_helper
1355 (pg, s, bi, start, last_start, n_alloc - n_added);
1359 if (PREDICT_FALSE (n_added < n_alloc))
1360 tail = clib_fifo_advance_tail (bi->buffer_fifo, n_added - n_alloc);
1365 /* Verify that pkts in the fifo are properly allocated */
1368 return n_in_fifo + n_added;
1378 /* Use pre data for packet data. */
1379 vlib_buffer_t buffer;
1383 format_pg_input_trace (u8 * s, va_list * va)
1385 vlib_main_t *vm = va_arg (*va, vlib_main_t *);
1386 CLIB_UNUSED (vlib_node_t * node) = va_arg (*va, vlib_node_t *);
1387 pg_input_trace_t *t = va_arg (*va, pg_input_trace_t *);
1388 pg_main_t *pg = &pg_main;
1389 pg_stream_t *stream;
1391 uword indent = format_get_indent (s);
1394 if (!pool_is_free_index (pg->streams, t->stream_index))
1395 stream = pool_elt_at_index (pg->streams, t->stream_index);
1398 s = format (s, "stream %v", pg->streams[t->stream_index].name);
1400 s = format (s, "stream %d", t->stream_index);
1402 s = format (s, ", %d bytes", t->packet_length);
1403 s = format (s, ", %d sw_if_index", t->sw_if_index);
1405 s = format (s, "\n%U%U",
1406 format_white_space, indent, format_vlib_buffer, &t->buffer);
1408 s = format (s, "\n%U", format_white_space, indent);
1412 n = vlib_get_node (vm, stream->node_index);
1414 if (n && n->format_buffer)
1415 s = format (s, "%U", n->format_buffer,
1416 t->buffer.pre_data, sizeof (t->buffer.pre_data));
1418 s = format (s, "%U",
1419 format_hex_bytes, t->buffer.pre_data,
1420 ARRAY_LEN (t->buffer.pre_data));
1425 pg_input_trace (pg_main_t * pg,
1426 vlib_node_runtime_t * node,
1427 pg_stream_t * s, u32 * buffers, u32 n_buffers)
1429 vlib_main_t *vm = vlib_get_main ();
1430 u32 *b, n_left, stream_index, next_index;
1434 stream_index = s - pg->streams;
1435 next_index = s->next_index;
1440 vlib_buffer_t *b0, *b1;
1441 pg_input_trace_t *t0, *t1;
1448 b0 = vlib_get_buffer (vm, bi0);
1449 b1 = vlib_get_buffer (vm, bi1);
1451 vlib_trace_buffer (vm, node, next_index, b0, /* follow_chain */ 1);
1452 vlib_trace_buffer (vm, node, next_index, b1, /* follow_chain */ 1);
1454 t0 = vlib_add_trace (vm, node, b0, sizeof (t0[0]));
1455 t1 = vlib_add_trace (vm, node, b1, sizeof (t1[0]));
1457 t0->stream_index = stream_index;
1458 t1->stream_index = stream_index;
1460 t0->packet_length = vlib_buffer_length_in_chain (vm, b0);
1461 t1->packet_length = vlib_buffer_length_in_chain (vm, b1);
1463 t0->sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX];
1464 t1->sw_if_index = vnet_buffer (b1)->sw_if_index[VLIB_RX];
1466 clib_memcpy (&t0->buffer, b0, sizeof (b0[0]) - sizeof (b0->pre_data));
1467 clib_memcpy (&t1->buffer, b1, sizeof (b1[0]) - sizeof (b1->pre_data));
1469 clib_memcpy (t0->buffer.pre_data, b0->data,
1470 sizeof (t0->buffer.pre_data));
1471 clib_memcpy (t1->buffer.pre_data, b1->data,
1472 sizeof (t1->buffer.pre_data));
1479 pg_input_trace_t *t0;
1485 b0 = vlib_get_buffer (vm, bi0);
1487 vlib_trace_buffer (vm, node, next_index, b0, /* follow_chain */ 1);
1488 t0 = vlib_add_trace (vm, node, b0, sizeof (t0[0]));
1490 t0->stream_index = stream_index;
1491 t0->packet_length = vlib_buffer_length_in_chain (vm, b0);
1492 t0->sw_if_index = vnet_buffer (b0)->sw_if_index[VLIB_RX];
1493 clib_memcpy (&t0->buffer, b0, sizeof (b0[0]) - sizeof (b0->pre_data));
1494 clib_memcpy (t0->buffer.pre_data, b0->data,
1495 sizeof (t0->buffer.pre_data));
1500 pg_generate_packets (vlib_node_runtime_t * node,
1502 pg_stream_t * s, uword n_packets_to_generate)
1504 vlib_main_t *vm = vlib_get_main ();
1505 u32 *to_next, n_this_frame, n_left, n_trace, n_packets_in_fifo;
1506 uword n_packets_generated;
1507 pg_buffer_index_t *bi, *bi0;
1508 u32 next_index = s->next_index;
1509 vnet_feature_main_t *fm = &feature_main;
1510 vnet_feature_config_main_t *cm;
1511 u8 feature_arc_index = fm->device_input_feature_arc_index;
1512 cm = &fm->feature_config_mains[feature_arc_index];
1513 u32 current_config_index = ~(u32) 0;
1516 bi0 = s->buffer_indices;
1518 n_packets_in_fifo = pg_stream_fill (pg, s, n_packets_to_generate);
1519 n_packets_to_generate = clib_min (n_packets_in_fifo, n_packets_to_generate);
1520 n_packets_generated = 0;
1523 (vnet_have_features (feature_arc_index, s->sw_if_index[VLIB_RX])))
1525 current_config_index =
1526 vec_elt (cm->config_index_by_sw_if_index, s->sw_if_index[VLIB_RX]);
1527 vnet_get_config_data (&cm->config_main, ¤t_config_index,
1531 while (n_packets_to_generate > 0)
1533 u32 *head, *start, *end;
1535 vlib_get_next_frame (vm, node, next_index, to_next, n_left);
1537 n_this_frame = n_packets_to_generate;
1538 if (n_this_frame > n_left)
1539 n_this_frame = n_left;
1541 start = bi0->buffer_fifo;
1542 end = clib_fifo_end (bi0->buffer_fifo);
1543 head = clib_fifo_head (bi0->buffer_fifo);
1545 if (head + n_this_frame <= end)
1546 vlib_copy_buffers (to_next, head, n_this_frame);
1550 vlib_copy_buffers (to_next + 0, head, n);
1551 vlib_copy_buffers (to_next + n, start, n_this_frame - n);
1554 vec_foreach (bi, s->buffer_indices)
1555 clib_fifo_advance_head (bi->buffer_fifo, n_this_frame);
1557 if (current_config_index != ~(u32) 0)
1558 for (i = 0; i < n_this_frame; i++)
1561 b = vlib_get_buffer (vm, to_next[i]);
1562 vnet_buffer (b)->device_input_feat.saved_next_index =
1564 vnet_buffer (b)->device_input_feat.buffer_advance = 0;
1565 b->current_config_index = current_config_index;
1566 b->feature_arc_index = feature_arc_index;
1569 n_trace = vlib_get_trace_count (vm, node);
1572 u32 n = clib_min (n_trace, n_this_frame);
1573 pg_input_trace (pg, node, s, to_next, n);
1574 vlib_set_trace_count (vm, node, n_trace - n);
1576 n_packets_to_generate -= n_this_frame;
1577 n_packets_generated += n_this_frame;
1578 n_left -= n_this_frame;
1579 vlib_put_next_frame (vm, node, next_index, n_left);
1582 return n_packets_generated;
1586 pg_input_stream (vlib_node_runtime_t * node, pg_main_t * pg, pg_stream_t * s)
1588 vlib_main_t *vm = vlib_get_main ();
1592 if (s->n_packets_limit > 0 && s->n_packets_generated >= s->n_packets_limit)
1594 pg_stream_enable_disable (pg, s, /* want_enabled */ 0);
1598 /* Apply rate limit. */
1599 time_now = vlib_time_now (vm);
1600 if (s->time_last_generate == 0)
1601 s->time_last_generate = time_now;
1603 dt = time_now - s->time_last_generate;
1604 s->time_last_generate = time_now;
1606 n_packets = VLIB_FRAME_SIZE;
1607 if (s->rate_packets_per_second > 0)
1609 s->packet_accumulator += dt * s->rate_packets_per_second;
1610 n_packets = s->packet_accumulator;
1612 /* Never allow accumulator to grow if we get behind. */
1613 s->packet_accumulator -= n_packets;
1616 /* Apply fixed limit. */
1617 if (s->n_packets_limit > 0
1618 && s->n_packets_generated + n_packets > s->n_packets_limit)
1619 n_packets = s->n_packets_limit - s->n_packets_generated;
1621 /* Generate up to one frame's worth of packets. */
1622 if (n_packets > VLIB_FRAME_SIZE)
1623 n_packets = VLIB_FRAME_SIZE;
1626 n_packets = pg_generate_packets (node, pg, s, n_packets);
1628 s->n_packets_generated += n_packets;
1634 pg_input (vlib_main_t * vm, vlib_node_runtime_t * node, vlib_frame_t * frame)
1637 pg_main_t *pg = &pg_main;
1638 uword n_packets = 0;
1639 u32 worker_index = 0;
1641 if (vlib_num_workers ())
1642 worker_index = vlib_get_current_worker_index ();
1645 clib_bitmap_foreach (i, pg->enabled_streams[worker_index], ({
1646 pg_stream_t *s = vec_elt_at_index (pg->streams, i);
1647 n_packets += pg_input_stream (node, pg, s);
1655 VLIB_REGISTER_NODE (pg_input_node) = {
1656 .function = pg_input,
1658 .sibling_of = "device-input",
1659 .type = VLIB_NODE_TYPE_INPUT,
1661 .format_trace = format_pg_input_trace,
1663 /* Input node will be left disabled until a stream is active. */
1664 .state = VLIB_NODE_STATE_DISABLED,
1669 * fd.io coding-style-patch-verification: ON
1672 * eval: (c-set-style "gnu")