next = nexts;
rx_comp = &rxq->rx_comp[comp_ring->next];
- while (PREDICT_TRUE (n_rx_packets < VLIB_FRAME_SIZE) &&
- (comp_ring->gen == (rx_comp->flags & VMXNET3_RXCF_GEN)))
+ while (PREDICT_TRUE ((n_rx_packets < VLIB_FRAME_SIZE) &&
+ (comp_ring->gen ==
+ (rx_comp->flags & VMXNET3_RXCF_GEN))))
{
vlib_buffer_t *b0;
u32 bi0;
u16 n_left = frame->n_vectors;
vmxnet3_txq_t *txq;
u32 thread_index = vm->thread_index;
- u16 qid = thread_index;
+ u16 qid = thread_index, produce;
if (PREDICT_FALSE (!(vd->flags & VMXNET3_DEVICE_F_LINK_UP)))
{
vmxnet3_txq_release (vm, vd, txq);
- while (n_left)
+ produce = txq->tx_ring.produce;
+ while (PREDICT_TRUE (n_left))
{
u16 space_needed = 1, i;
vlib_buffer_t *b;
* Device can start reading the packet
*/
txq->tx_desc[first_idx].flags[0] ^= VMXNET3_TXF_GEN;
- vmxnet3_reg_write_inline (vd, 0, VMXNET3_REG_TXPROD,
- txq->tx_ring.produce);
buffers++;
n_left--;
}
+ if (PREDICT_TRUE (produce != txq->tx_ring.produce))
+ vmxnet3_reg_write_inline (vd, 0, VMXNET3_REG_TXPROD,
+ txq->tx_ring.produce);
+
clib_spinlock_unlock_if_init (&txq->lock);
return (frame->n_vectors - n_left);
vmxnet3_rx_desc *rxd;
u16 n_refill, n_alloc;
vmxnet3_rx_ring *ring;
+ vmxnet3_queues *q;
ring = &rxq->rx_ring[0];
n_refill = rxq->size - ring->fill;
n_alloc--;
}
- vmxnet3_reg_write_inline (vd, 0, VMXNET3_REG_RXPROD, ring->produce);
+ q = &vd->dma->queues;
+ if (PREDICT_FALSE (q->rx.ctrl.update_prod))
+ vmxnet3_reg_write_inline (vd, 0, VMXNET3_REG_RXPROD, ring->produce);
return 0;
}
vmxnet3_rx_desc *rxd;
u16 n_refill, n_alloc;
vmxnet3_rx_ring *ring;
+ vmxnet3_queues *q;
ring = &rxq->rx_ring[1];
n_refill = rxq->size - ring->fill;
n_alloc--;
}
- vmxnet3_reg_write_inline (vd, 0, VMXNET3_REG_RXPROD2, ring->produce);
+ q = &vd->dma->queues;
+ if (PREDICT_FALSE (q->rx.ctrl.update_prod))
+ vmxnet3_reg_write_inline (vd, 0, VMXNET3_REG_RXPROD2, ring->produce);
return 0;
}