X-Git-Url: https://gerrit.fd.io/r/gitweb?a=blobdiff_plain;f=src%2Fplugins%2Favf%2Fdevice.c;h=248e3c173bd41b1e6c5963d42451ab3eec8d5129;hb=34931eb47124ef78a3aa9575930e848a78c9618b;hp=6bf1fe05f05eac7ace75a488c87788063a9f6385;hpb=5100aa9cb9e7acff35fa3bfde8aa95b5ace60344;p=vpp.git diff --git a/src/plugins/avf/device.c b/src/plugins/avf/device.c index 6bf1fe05f05..248e3c173bd 100644 --- a/src/plugins/avf/device.c +++ b/src/plugins/avf/device.c @@ -16,6 +16,7 @@ */ #include +#include #include #include #include @@ -106,7 +107,7 @@ avf_aq_desc_enq (vlib_main_t * vm, avf_device_t * ad, avf_aq_desc_t * dt, int n_retry = 5; d = &ad->atq[ad->atq_next_slot]; - clib_memcpy (d, dt, sizeof (avf_aq_desc_t)); + clib_memcpy_fast (d, dt, sizeof (avf_aq_desc_t)); d->flags |= AVF_AQ_F_RD | AVF_AQ_F_SI; if (len) d->datalen = len; @@ -116,13 +117,13 @@ avf_aq_desc_enq (vlib_main_t * vm, avf_device_t * ad, avf_aq_desc_t * dt, pa = ad->atq_bufs_pa + ad->atq_next_slot * AVF_MBOX_BUF_SZ; d->addr_hi = (u32) (pa >> 32); d->addr_lo = (u32) pa; - clib_memcpy (ad->atq_bufs + ad->atq_next_slot * AVF_MBOX_BUF_SZ, data, - len); + clib_memcpy_fast (ad->atq_bufs + ad->atq_next_slot * AVF_MBOX_BUF_SZ, + data, len); d->flags |= AVF_AQ_F_BUF; } if (ad->flags & AVF_DEVICE_F_ELOG) - clib_memcpy (&dc, d, sizeof (avf_aq_desc_t)); + clib_memcpy_fast (&dc, d, sizeof (avf_aq_desc_t)); CLIB_MEMORY_BARRIER (); vlib_log_debug (am->log_class, "%U", format_hexdump, data, len); @@ -144,7 +145,7 @@ retry: goto retry; } - clib_memcpy (dt, d, sizeof (avf_aq_desc_t)); + clib_memcpy_fast (dt, d, sizeof (avf_aq_desc_t)); if (d->flags & AVF_AQ_F_ERR) return clib_error_return (0, "adminq enqueue error [opcode 0x%x, retval " "%d]", d->opcode, d->retval); @@ -223,9 +224,14 @@ avf_rxq_init (vlib_main_t * vm, avf_device_t * ad, u16 qid, u16 rxq_size) rxq = vec_elt_at_index (ad->rxqs, qid); rxq->size = rxq_size; rxq->next = 0; - rxq->descs = vlib_physmem_alloc_aligned (vm, rxq->size * - sizeof (avf_rx_desc_t), - 2 * CLIB_CACHE_LINE_BYTES); + rxq->descs = vlib_physmem_alloc_aligned_on_numa (vm, rxq->size * + sizeof (avf_rx_desc_t), + 2 * CLIB_CACHE_LINE_BYTES, + ad->numa_node); + + rxq->buffer_pool_index = + vlib_buffer_pool_get_default_for_numa (vm, ad->numa_node); + if (rxq->descs == 0) return vlib_physmem_last_error (vm); @@ -236,7 +242,8 @@ avf_rxq_init (vlib_main_t * vm, avf_device_t * ad, u16 qid, u16 rxq_size) vec_validate_aligned (rxq->bufs, rxq->size, CLIB_CACHE_LINE_BYTES); rxq->qrx_tail = ad->bar0 + AVF_QRX_TAIL (qid); - n_alloc = vlib_buffer_alloc (vm, rxq->bufs, rxq->size - 8); + n_alloc = vlib_buffer_alloc_from_pool (vm, rxq->bufs, rxq->size - 8, + rxq->buffer_pool_index); if (n_alloc == 0) return clib_error_return (0, "buffer allocation error"); @@ -277,9 +284,10 @@ avf_txq_init (vlib_main_t * vm, avf_device_t * ad, u16 qid, u16 txq_size) txq = vec_elt_at_index (ad->txqs, qid); txq->size = txq_size; txq->next = 0; - txq->descs = vlib_physmem_alloc_aligned (vm, txq->size * - sizeof (avf_tx_desc_t), - 2 * CLIB_CACHE_LINE_BYTES); + txq->descs = vlib_physmem_alloc_aligned_on_numa (vm, txq->size * + sizeof (avf_tx_desc_t), + 2 * CLIB_CACHE_LINE_BYTES, + ad->numa_node); if (txq->descs == 0) return vlib_physmem_last_error (vm); @@ -289,6 +297,9 @@ avf_txq_init (vlib_main_t * vm, avf_device_t * ad, u16 qid, u16 txq_size) vec_validate_aligned (txq->bufs, txq->size, CLIB_CACHE_LINE_BYTES); txq->qtx_tail = ad->bar0 + AVF_QTX_TAIL (qid); + /* initialize ring of pending RS slots */ + clib_ring_new_aligned (txq->rs_slots, 32, CLIB_CACHE_LINE_BYTES); + ad->n_tx_queues = clib_min (ad->num_queue_pairs, qid + 1); return 0; } @@ -398,7 +409,7 @@ retry: return clib_error_return (0, "event message error"); vec_add2 (ad->events, e, 1); - clib_memcpy (e, buf, sizeof (virtchnl_pf_event_t)); + clib_memcpy_fast (e, buf, sizeof (virtchnl_pf_event_t)); avf_arq_slot_init (ad, ad->arq_next_slot); ad->arq_next_slot++; n_retry = 5; @@ -425,7 +436,7 @@ retry: if (d->flags & AVF_AQ_F_BUF) { void *buf = ad->arq_bufs + ad->arq_next_slot * AVF_MBOX_BUF_SZ; - clib_memcpy (out, buf, out_len); + clib_memcpy_fast (out, buf, out_len); } avf_arq_slot_init (ad, ad->arq_next_slot); @@ -578,12 +589,12 @@ avf_op_config_vsi_queues (vlib_main_t * vm, avf_device_t * ad) rxq->vsi_id = ad->vsi_id; rxq->queue_id = i; - rxq->max_pkt_size = 1518; + rxq->max_pkt_size = ETHERNET_MAX_PACKET_BYTES; if (i < vec_len (ad->rxqs)) { avf_rxq_t *q = vec_elt_at_index (ad->rxqs, i); rxq->ring_len = q->size; - rxq->databuffer_size = VLIB_BUFFER_DEFAULT_FREE_LIST_BYTES; + rxq->databuffer_size = vlib_buffer_get_default_data_size (vm); rxq->dma_ring_addr = avf_dma_addr (vm, ad, (void *) q->descs); avf_reg_write (ad, AVF_QRX_TAIL (i), q->size - 1); } @@ -637,7 +648,7 @@ avf_op_add_eth_addr (vlib_main_t * vm, avf_device_t * ad, u8 count, u8 * macs) al->vsi_id = ad->vsi_id; al->num_elements = count; for (i = 0; i < count; i++) - clib_memcpy (&al->list[i].addr, macs + i * 6, 6); + clib_memcpy_fast (&al->list[i].addr, macs + i * 6, 6); return avf_send_to_pf (vm, ad, VIRTCHNL_OP_ADD_ETH_ADDR, msg, msg_len, 0, 0); } @@ -646,14 +657,19 @@ clib_error_t * avf_op_enable_queues (vlib_main_t * vm, avf_device_t * ad, u32 rx, u32 tx) { virtchnl_queue_select_t qs = { 0 }; - int i; + int i = 0; qs.vsi_id = ad->vsi_id; qs.rx_queues = rx; qs.tx_queues = tx; - for (i = 0; i < ad->n_rx_queues; i++) + while (rx) { - avf_rxq_t *rxq = vec_elt_at_index (ad->rxqs, i); - avf_reg_write (ad, AVF_QRX_TAIL (i), rxq->n_enqueued); + if (rx & (1 << i)) + { + avf_rxq_t *rxq = vec_elt_at_index (ad->rxqs, i); + avf_reg_write (ad, AVF_QRX_TAIL (i), rxq->n_enqueued); + rx &= ~(1 << i); + } + i++; } return avf_send_to_pf (vm, ad, VIRTCHNL_OP_ENABLE_QUEUES, &qs, sizeof (virtchnl_queue_select_t), 0, 0); @@ -788,7 +804,7 @@ avf_device_init (vlib_main_t * vm, avf_main_t * am, avf_device_t * ad, ad->rss_key_size = res.rss_key_size; ad->rss_lut_size = res.rss_lut_size; - clib_memcpy (ad->hwaddr, res.vsi_res[0].default_mac_addr, 6); + clib_memcpy_fast (ad->hwaddr, res.vsi_res[0].default_mac_addr, 6); /* * Disable VLAN stripping @@ -843,10 +859,8 @@ avf_device_init (vlib_main_t * vm, avf_main_t * am, avf_device_t * ad, if ((error = avf_op_add_eth_addr (vm, ad, 1, ad->hwaddr))) return error; - if ((error = avf_op_enable_queues (vm, ad, ad->n_rx_queues, 0))) - return error; - - if ((error = avf_op_enable_queues (vm, ad, 0, ad->n_tx_queues))) + if ((error = avf_op_enable_queues (vm, ad, pow2_mask (ad->n_rx_queues), + pow2_mask (ad->n_tx_queues)))) return error; ad->flags |= AVF_DEVICE_F_INITIALIZED; @@ -1166,9 +1180,11 @@ avf_delete_if (vlib_main_t * vm, avf_device_t * ad) txq->n_enqueued); } vec_free (txq->bufs); + clib_ring_free (txq->rs_slots); } /* *INDENT-ON* */ vec_free (ad->txqs); + vec_free (ad->name); clib_error_free (ad->error); clib_memset (ad, 0, sizeof (*ad)); @@ -1201,6 +1217,7 @@ avf_create_if (vlib_main_t * vm, avf_create_if_args_t * args) pool_get (am->devices, ad); ad->dev_instance = ad - am->devices; ad->per_interface_next_index = ~0; + ad->name = vec_dup (args->name); if (args->enable_elog) ad->flags |= AVF_DEVICE_F_ELOG; @@ -1216,6 +1233,7 @@ avf_create_if (vlib_main_t * vm, avf_create_if_args_t * args) return; } ad->pci_dev_handle = h; + ad->numa_node = vlib_pci_get_numa_node (vm, h); vlib_pci_set_private_data (vm, h, ad->dev_instance); @@ -1236,8 +1254,11 @@ avf_create_if (vlib_main_t * vm, avf_create_if_args_t * args) if ((error = vlib_pci_enable_msix_irq (vm, h, 0, 2))) goto error; - if (!(ad->atq = vlib_physmem_alloc (vm, sizeof (avf_aq_desc_t) * - AVF_MBOX_LEN))) + ad->atq = vlib_physmem_alloc_aligned_on_numa (vm, sizeof (avf_aq_desc_t) * + AVF_MBOX_LEN, + CLIB_CACHE_LINE_BYTES, + ad->numa_node); + if (ad->atq == 0) { error = vlib_physmem_last_error (vm); goto error; @@ -1246,8 +1267,11 @@ avf_create_if (vlib_main_t * vm, avf_create_if_args_t * args) if ((error = vlib_pci_map_dma (vm, h, ad->atq))) goto error; - if (!(ad->arq = vlib_physmem_alloc (vm, sizeof (avf_aq_desc_t) * - AVF_MBOX_LEN))) + ad->arq = vlib_physmem_alloc_aligned_on_numa (vm, sizeof (avf_aq_desc_t) * + AVF_MBOX_LEN, + CLIB_CACHE_LINE_BYTES, + ad->numa_node); + if (ad->arq == 0) { error = vlib_physmem_last_error (vm); goto error; @@ -1256,8 +1280,11 @@ avf_create_if (vlib_main_t * vm, avf_create_if_args_t * args) if ((error = vlib_pci_map_dma (vm, h, ad->arq))) goto error; - if (!(ad->atq_bufs = vlib_physmem_alloc (vm, AVF_MBOX_BUF_SZ * - AVF_MBOX_LEN))) + ad->atq_bufs = vlib_physmem_alloc_aligned_on_numa (vm, AVF_MBOX_BUF_SZ * + AVF_MBOX_LEN, + CLIB_CACHE_LINE_BYTES, + ad->numa_node); + if (ad->atq_bufs == 0) { error = vlib_physmem_last_error (vm); goto error; @@ -1266,8 +1293,11 @@ avf_create_if (vlib_main_t * vm, avf_create_if_args_t * args) if ((error = vlib_pci_map_dma (vm, h, ad->atq_bufs))) goto error; - if (!(ad->arq_bufs = vlib_physmem_alloc (vm, AVF_MBOX_BUF_SZ * - AVF_MBOX_LEN))) + ad->arq_bufs = vlib_physmem_alloc_aligned_on_numa (vm, AVF_MBOX_BUF_SZ * + AVF_MBOX_LEN, + CLIB_CACHE_LINE_BYTES, + ad->numa_node); + if (ad->arq_bufs == 0) { error = vlib_physmem_last_error (vm); goto error; @@ -1405,7 +1435,6 @@ avf_init (vlib_main_t * vm) avf_main_t *am = &avf_main; clib_error_t *error; vlib_thread_main_t *tm = vlib_get_thread_main (); - int i; if ((error = vlib_call_init_function (vm, pci_bus_init))) return error; @@ -1413,30 +1442,6 @@ avf_init (vlib_main_t * vm) vec_validate_aligned (am->per_thread_data, tm->n_vlib_mains - 1, CLIB_CACHE_LINE_BYTES); - /* initialize ptype based loopup table */ - vec_validate_aligned (am->ptypes, 255, CLIB_CACHE_LINE_BYTES); - - /* *INDENT-OFF* */ - vec_foreach_index (i, am->ptypes) - { - avf_ptype_t *p = vec_elt_at_index (am->ptypes, i); - if ((i >= 22) && (i <= 87)) - { - p->next_node = VNET_DEVICE_INPUT_NEXT_IP4_NCS_INPUT; - p->flags = VNET_BUFFER_F_IS_IP4; - } - else if ((i >= 88) && (i <= 153)) - { - p->next_node = VNET_DEVICE_INPUT_NEXT_IP6_INPUT; - p->flags = VNET_BUFFER_F_IS_IP6; - } - else - p->next_node = VNET_DEVICE_INPUT_NEXT_ETHERNET_INPUT; - p->buffer_advance = device_input_next_node_advance[p->next_node]; - p->flags |= VLIB_BUFFER_TOTAL_LENGTH_VALID; - } - /* *INDENT-ON* */ - am->log_class = vlib_log_register_class ("avf_plugin", 0); vlib_log_debug (am->log_class, "initialized");