u8 qid = thread_index;
avf_txq_t *txq = vec_elt_at_index (ad->txqs, qid % ad->num_queue_pairs);
avf_tx_desc_t *d0, *d1, *d2, *d3;
- u32 *buffers = vlib_frame_args (frame);
+ u32 *buffers = vlib_frame_vector_args (frame);
u32 bi0, bi1, bi2, bi3;
- u16 n_left = frame->n_vectors;
+ u16 n_left, n_left_to_send, n_in_batch;
vlib_buffer_t *b0, *b1, *b2, *b3;
+ u16 next;
+ u16 n_retry = 5;
u16 mask = txq->size - 1;
u64 bits = (AVF_TXQ_DESC_CMD_EOP | AVF_TXQ_DESC_CMD_RS |
AVF_TXQ_DESC_CMD_RSV);
clib_spinlock_lock_if_init (&txq->lock);
- /* release cosumed bufs */
+ n_left_to_send = frame->n_vectors;
+ next = txq->next;
+
+retry:
+ /* release consumed bufs */
if (txq->n_enqueued)
{
u16 first, slot, n_free = 0;
- first = slot = (txq->next - txq->n_enqueued) & mask;
+ first = slot = (next - txq->n_enqueued) & mask;
d0 = txq->descs + slot;
while (n_free < txq->n_enqueued && avf_tx_desc_get_dtyp (d0) == 0x0F)
{
}
}
+ n_in_batch = clib_min (n_left_to_send, txq->size - txq->n_enqueued - 8);
+ n_left = n_in_batch;
+
while (n_left >= 8)
{
u16 slot0, slot1, slot2, slot3;
vlib_prefetch_buffer_with_index (vm, buffers[6], LOAD);
vlib_prefetch_buffer_with_index (vm, buffers[7], LOAD);
- slot0 = txq->next;
- slot1 = (txq->next + 1) & mask;
- slot2 = (txq->next + 2) & mask;
- slot3 = (txq->next + 3) & mask;
+ slot0 = next;
+ slot1 = (next + 1) & mask;
+ slot2 = (next + 2) & mask;
+ slot3 = (next + 3) & mask;
d0 = txq->descs + slot0;
d1 = txq->descs + slot1;
b2 = vlib_get_buffer (vm, bi2);
b3 = vlib_get_buffer (vm, bi3);
-#if 0
- d->qword[0] = vlib_get_buffer_data_physical_address (vm, bi0) +
- b0->current_data;
-#else
- d0->qword[0] = pointer_to_uword (b0->data) + b0->current_data;
- d1->qword[0] = pointer_to_uword (b1->data) + b1->current_data;
- d2->qword[0] = pointer_to_uword (b2->data) + b2->current_data;
- d3->qword[0] = pointer_to_uword (b3->data) + b3->current_data;
+ if (ad->flags & AVF_DEVICE_F_VA_DMA)
+ {
+ d0->qword[0] = vlib_buffer_get_current_va (b0);
+ d1->qword[0] = vlib_buffer_get_current_va (b1);
+ d2->qword[0] = vlib_buffer_get_current_va (b2);
+ d3->qword[0] = vlib_buffer_get_current_va (b3);
+ }
+ else
+ {
+ d0->qword[0] = vlib_buffer_get_current_pa (vm, b0);
+ d1->qword[0] = vlib_buffer_get_current_pa (vm, b1);
+ d2->qword[0] = vlib_buffer_get_current_pa (vm, b2);
+ d3->qword[0] = vlib_buffer_get_current_pa (vm, b3);
+ }
-#endif
d0->qword[1] = ((u64) b0->current_length) << 34 | bits;
d1->qword[1] = ((u64) b1->current_length) << 34 | bits;
d2->qword[1] = ((u64) b2->current_length) << 34 | bits;
d3->qword[1] = ((u64) b3->current_length) << 34 | bits;
- txq->next = (txq->next + 4) & mask;
+ next = (next + 4) & mask;
txq->n_enqueued += 4;
buffers += 4;
n_left -= 4;
while (n_left)
{
- d0 = txq->descs + txq->next;
+ d0 = txq->descs + next;
bi0 = buffers[0];
- txq->bufs[txq->next] = bi0;
+ txq->bufs[next] = bi0;
b0 = vlib_get_buffer (vm, bi0);
-#if 0
- d->qword[0] = vlib_get_buffer_data_physical_address (vm, bi0) +
- b0->current_data;
-#else
- d0->qword[0] = pointer_to_uword (b0->data) + b0->current_data;
-#endif
+ if (ad->flags & AVF_DEVICE_F_VA_DMA)
+ d0->qword[0] = vlib_buffer_get_current_va (b0);
+ else
+ d0->qword[0] = vlib_buffer_get_current_pa (vm, b0);
+
d0->qword[1] = (((u64) b0->current_length) << 34) | bits;
- txq->next = (txq->next + 1) & mask;
+ next = (next + 1) & mask;
txq->n_enqueued++;
buffers++;
n_left--;
}
+
CLIB_MEMORY_BARRIER ();
- *(txq->qtx_tail) = txq->next;
+ *(txq->qtx_tail) = txq->next = next;
+
+ n_left_to_send -= n_in_batch;
+
+ if (n_left_to_send)
+ {
+ if (n_retry--)
+ goto retry;
+
+ vlib_buffer_free (vm, buffers, n_left_to_send);
+ vlib_error_count (vm, node->node_index,
+ AVF_TX_ERROR_NO_FREE_SLOTS, n_left_to_send);
+ }
clib_spinlock_unlock_if_init (&txq->lock);