hw/net/net_tx_pkt: Decouple interface from PCI

This allows to use the network packet abstractions even if PCI is not
used.

Signed-off-by: Akihiko Odaki <akihiko.odaki@daynix.com>
Signed-off-by: Jason Wang <jasowang@redhat.com>
This commit is contained in:
Akihiko Odaki 2023-05-23 11:42:53 +09:00 committed by Jason Wang
parent 163246e1ce
commit a51db58027
5 changed files with 54 additions and 53 deletions

View file

@ -746,7 +746,8 @@ e1000e_process_tx_desc(E1000ECore *core,
addr = le64_to_cpu(dp->buffer_addr); addr = le64_to_cpu(dp->buffer_addr);
if (!tx->skip_cp) { if (!tx->skip_cp) {
if (!net_tx_pkt_add_raw_fragment(tx->tx_pkt, addr, split_size)) { if (!net_tx_pkt_add_raw_fragment_pci(tx->tx_pkt, core->owner,
addr, split_size)) {
tx->skip_cp = true; tx->skip_cp = true;
} }
} }
@ -764,7 +765,7 @@ e1000e_process_tx_desc(E1000ECore *core,
} }
tx->skip_cp = false; tx->skip_cp = false;
net_tx_pkt_reset(tx->tx_pkt, core->owner); net_tx_pkt_reset(tx->tx_pkt, net_tx_pkt_unmap_frag_pci, core->owner);
tx->sum_needed = 0; tx->sum_needed = 0;
tx->cptse = 0; tx->cptse = 0;
@ -3421,7 +3422,7 @@ e1000e_core_pci_realize(E1000ECore *core,
qemu_add_vm_change_state_handler(e1000e_vm_state_change, core); qemu_add_vm_change_state_handler(e1000e_vm_state_change, core);
for (i = 0; i < E1000E_NUM_QUEUES; i++) { for (i = 0; i < E1000E_NUM_QUEUES; i++) {
net_tx_pkt_init(&core->tx[i].tx_pkt, core->owner, E1000E_MAX_TX_FRAGS); net_tx_pkt_init(&core->tx[i].tx_pkt, E1000E_MAX_TX_FRAGS);
} }
net_rx_pkt_init(&core->rx_pkt); net_rx_pkt_init(&core->rx_pkt);
@ -3446,7 +3447,8 @@ e1000e_core_pci_uninit(E1000ECore *core)
qemu_del_vm_change_state_handler(core->vmstate); qemu_del_vm_change_state_handler(core->vmstate);
for (i = 0; i < E1000E_NUM_QUEUES; i++) { for (i = 0; i < E1000E_NUM_QUEUES; i++) {
net_tx_pkt_reset(core->tx[i].tx_pkt, core->owner); net_tx_pkt_reset(core->tx[i].tx_pkt,
net_tx_pkt_unmap_frag_pci, core->owner);
net_tx_pkt_uninit(core->tx[i].tx_pkt); net_tx_pkt_uninit(core->tx[i].tx_pkt);
} }
@ -3571,7 +3573,8 @@ static void e1000e_reset(E1000ECore *core, bool sw)
e1000x_reset_mac_addr(core->owner_nic, core->mac, core->permanent_mac); e1000x_reset_mac_addr(core->owner_nic, core->mac, core->permanent_mac);
for (i = 0; i < ARRAY_SIZE(core->tx); i++) { for (i = 0; i < ARRAY_SIZE(core->tx); i++) {
net_tx_pkt_reset(core->tx[i].tx_pkt, core->owner); net_tx_pkt_reset(core->tx[i].tx_pkt,
net_tx_pkt_unmap_frag_pci, core->owner);
memset(&core->tx[i].props, 0, sizeof(core->tx[i].props)); memset(&core->tx[i].props, 0, sizeof(core->tx[i].props));
core->tx[i].skip_cp = false; core->tx[i].skip_cp = false;
} }

View file

@ -597,7 +597,8 @@ igb_process_tx_desc(IGBCore *core,
length = cmd_type_len & 0xFFFF; length = cmd_type_len & 0xFFFF;
if (!tx->skip_cp) { if (!tx->skip_cp) {
if (!net_tx_pkt_add_raw_fragment(tx->tx_pkt, buffer_addr, length)) { if (!net_tx_pkt_add_raw_fragment_pci(tx->tx_pkt, dev,
buffer_addr, length)) {
tx->skip_cp = true; tx->skip_cp = true;
} }
} }
@ -616,7 +617,7 @@ igb_process_tx_desc(IGBCore *core,
tx->first = true; tx->first = true;
tx->skip_cp = false; tx->skip_cp = false;
net_tx_pkt_reset(tx->tx_pkt, dev); net_tx_pkt_reset(tx->tx_pkt, net_tx_pkt_unmap_frag_pci, dev);
} }
} }
@ -842,8 +843,6 @@ igb_start_xmit(IGBCore *core, const IGB_TxRing *txr)
d = core->owner; d = core->owner;
} }
net_tx_pkt_reset(txr->tx->tx_pkt, d);
while (!igb_ring_empty(core, txi)) { while (!igb_ring_empty(core, txi)) {
base = igb_ring_head_descr(core, txi); base = igb_ring_head_descr(core, txi);
@ -861,6 +860,8 @@ igb_start_xmit(IGBCore *core, const IGB_TxRing *txr)
core->mac[EICR] |= eic; core->mac[EICR] |= eic;
igb_set_interrupt_cause(core, E1000_ICR_TXDW); igb_set_interrupt_cause(core, E1000_ICR_TXDW);
} }
net_tx_pkt_reset(txr->tx->tx_pkt, net_tx_pkt_unmap_frag_pci, d);
} }
static uint32_t static uint32_t
@ -3954,7 +3955,7 @@ igb_core_pci_realize(IGBCore *core,
core->vmstate = qemu_add_vm_change_state_handler(igb_vm_state_change, core); core->vmstate = qemu_add_vm_change_state_handler(igb_vm_state_change, core);
for (i = 0; i < IGB_NUM_QUEUES; i++) { for (i = 0; i < IGB_NUM_QUEUES; i++) {
net_tx_pkt_init(&core->tx[i].tx_pkt, NULL, E1000E_MAX_TX_FRAGS); net_tx_pkt_init(&core->tx[i].tx_pkt, E1000E_MAX_TX_FRAGS);
} }
net_rx_pkt_init(&core->rx_pkt); net_rx_pkt_init(&core->rx_pkt);
@ -3979,7 +3980,6 @@ igb_core_pci_uninit(IGBCore *core)
qemu_del_vm_change_state_handler(core->vmstate); qemu_del_vm_change_state_handler(core->vmstate);
for (i = 0; i < IGB_NUM_QUEUES; i++) { for (i = 0; i < IGB_NUM_QUEUES; i++) {
net_tx_pkt_reset(core->tx[i].tx_pkt, NULL);
net_tx_pkt_uninit(core->tx[i].tx_pkt); net_tx_pkt_uninit(core->tx[i].tx_pkt);
} }
@ -4158,7 +4158,6 @@ static void igb_reset(IGBCore *core, bool sw)
for (i = 0; i < ARRAY_SIZE(core->tx); i++) { for (i = 0; i < ARRAY_SIZE(core->tx); i++) {
tx = &core->tx[i]; tx = &core->tx[i];
net_tx_pkt_reset(tx->tx_pkt, NULL);
memset(tx->ctx, 0, sizeof(tx->ctx)); memset(tx->ctx, 0, sizeof(tx->ctx));
tx->first = true; tx->first = true;
tx->skip_cp = false; tx->skip_cp = false;

View file

@ -16,12 +16,12 @@
*/ */
#include "qemu/osdep.h" #include "qemu/osdep.h"
#include "net_tx_pkt.h"
#include "net/eth.h" #include "net/eth.h"
#include "net/checksum.h" #include "net/checksum.h"
#include "net/tap.h" #include "net/tap.h"
#include "net/net.h" #include "net/net.h"
#include "hw/pci/pci_device.h" #include "hw/pci/pci_device.h"
#include "net_tx_pkt.h"
enum { enum {
NET_TX_PKT_VHDR_FRAG = 0, NET_TX_PKT_VHDR_FRAG = 0,
@ -32,8 +32,6 @@ enum {
/* TX packet private context */ /* TX packet private context */
struct NetTxPkt { struct NetTxPkt {
PCIDevice *pci_dev;
struct virtio_net_hdr virt_hdr; struct virtio_net_hdr virt_hdr;
struct iovec *raw; struct iovec *raw;
@ -59,13 +57,10 @@ struct NetTxPkt {
uint8_t l4proto; uint8_t l4proto;
}; };
void net_tx_pkt_init(struct NetTxPkt **pkt, PCIDevice *pci_dev, void net_tx_pkt_init(struct NetTxPkt **pkt, uint32_t max_frags)
uint32_t max_frags)
{ {
struct NetTxPkt *p = g_malloc0(sizeof *p); struct NetTxPkt *p = g_malloc0(sizeof *p);
p->pci_dev = pci_dev;
p->vec = g_new(struct iovec, max_frags + NET_TX_PKT_PL_START_FRAG); p->vec = g_new(struct iovec, max_frags + NET_TX_PKT_PL_START_FRAG);
p->raw = g_new(struct iovec, max_frags); p->raw = g_new(struct iovec, max_frags);
@ -384,8 +379,7 @@ void net_tx_pkt_setup_vlan_header_ex(struct NetTxPkt *pkt,
} }
} }
static bool net_tx_pkt_add_raw_fragment_common(struct NetTxPkt *pkt, bool net_tx_pkt_add_raw_fragment(struct NetTxPkt *pkt, void *base, size_t len)
void *base, size_t len)
{ {
struct iovec *ventry; struct iovec *ventry;
assert(pkt); assert(pkt);
@ -433,7 +427,8 @@ void net_tx_pkt_dump(struct NetTxPkt *pkt)
#endif #endif
} }
void net_tx_pkt_reset(struct NetTxPkt *pkt, PCIDevice *pci_dev) void net_tx_pkt_reset(struct NetTxPkt *pkt,
NetTxPktFreeFrag callback, void *context)
{ {
int i; int i;
@ -453,12 +448,9 @@ void net_tx_pkt_reset(struct NetTxPkt *pkt, PCIDevice *pci_dev)
assert(pkt->raw); assert(pkt->raw);
for (i = 0; i < pkt->raw_frags; i++) { for (i = 0; i < pkt->raw_frags; i++) {
assert(pkt->raw[i].iov_base); assert(pkt->raw[i].iov_base);
net_tx_pkt_unmap_frag_pci(pkt->pci_dev, callback(context, pkt->raw[i].iov_base, pkt->raw[i].iov_len);
pkt->raw[i].iov_base,
pkt->raw[i].iov_len);
} }
} }
pkt->pci_dev = pci_dev;
pkt->raw_frags = 0; pkt->raw_frags = 0;
pkt->hdr_len = 0; pkt->hdr_len = 0;
@ -470,19 +462,17 @@ void net_tx_pkt_unmap_frag_pci(void *context, void *base, size_t len)
pci_dma_unmap(context, base, len, DMA_DIRECTION_TO_DEVICE, 0); pci_dma_unmap(context, base, len, DMA_DIRECTION_TO_DEVICE, 0);
} }
bool net_tx_pkt_add_raw_fragment(struct NetTxPkt *pkt, hwaddr pa, bool net_tx_pkt_add_raw_fragment_pci(struct NetTxPkt *pkt, PCIDevice *pci_dev,
size_t len) dma_addr_t pa, size_t len)
{ {
dma_addr_t mapped_len = len; dma_addr_t mapped_len = len;
void *base = pci_dma_map(pkt->pci_dev, pa, &mapped_len, void *base = pci_dma_map(pci_dev, pa, &mapped_len, DMA_DIRECTION_TO_DEVICE);
DMA_DIRECTION_TO_DEVICE);
if (!base) { if (!base) {
return false; return false;
} }
if (mapped_len != len || if (mapped_len != len || !net_tx_pkt_add_raw_fragment(pkt, base, len)) {
!net_tx_pkt_add_raw_fragment_common(pkt, base, len)) { net_tx_pkt_unmap_frag_pci(pci_dev, base, mapped_len);
net_tx_pkt_unmap_frag_pci(pkt->pci_dev, base, mapped_len);
return false; return false;
} }
@ -710,7 +700,7 @@ static void net_tx_pkt_udp_fragment_fix(struct NetTxPkt *pkt,
} }
static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt, static bool net_tx_pkt_do_sw_fragmentation(struct NetTxPkt *pkt,
NetTxPktCallback callback, NetTxPktSend callback,
void *context) void *context)
{ {
uint8_t gso_type = pkt->virt_hdr.gso_type & ~VIRTIO_NET_HDR_GSO_ECN; uint8_t gso_type = pkt->virt_hdr.gso_type & ~VIRTIO_NET_HDR_GSO_ECN;
@ -807,7 +797,7 @@ bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc)
} }
bool net_tx_pkt_send_custom(struct NetTxPkt *pkt, bool offload, bool net_tx_pkt_send_custom(struct NetTxPkt *pkt, bool offload,
NetTxPktCallback callback, void *context) NetTxPktSend callback, void *context)
{ {
assert(pkt); assert(pkt);

View file

@ -26,17 +26,16 @@
struct NetTxPkt; struct NetTxPkt;
typedef void (* NetTxPktCallback)(void *, const struct iovec *, int, const struct iovec *, int); typedef void (*NetTxPktFreeFrag)(void *, void *, size_t);
typedef void (*NetTxPktSend)(void *, const struct iovec *, int, const struct iovec *, int);
/** /**
* Init function for tx packet functionality * Init function for tx packet functionality
* *
* @pkt: packet pointer * @pkt: packet pointer
* @pci_dev: PCI device processing this packet
* @max_frags: max tx ip fragments * @max_frags: max tx ip fragments
*/ */
void net_tx_pkt_init(struct NetTxPkt **pkt, PCIDevice *pci_dev, void net_tx_pkt_init(struct NetTxPkt **pkt, uint32_t max_frags);
uint32_t max_frags);
/** /**
* Clean all tx packet resources. * Clean all tx packet resources.
@ -95,12 +94,11 @@ net_tx_pkt_setup_vlan_header(struct NetTxPkt *pkt, uint16_t vlan)
* populate data fragment into pkt context. * populate data fragment into pkt context.
* *
* @pkt: packet * @pkt: packet
* @pa: physical address of fragment * @base: pointer to fragment
* @len: length of fragment * @len: length of fragment
* *
*/ */
bool net_tx_pkt_add_raw_fragment(struct NetTxPkt *pkt, hwaddr pa, bool net_tx_pkt_add_raw_fragment(struct NetTxPkt *pkt, void *base, size_t len);
size_t len);
/** /**
* Fix ip header fields and calculate IP header and pseudo header checksums. * Fix ip header fields and calculate IP header and pseudo header checksums.
@ -148,10 +146,11 @@ void net_tx_pkt_dump(struct NetTxPkt *pkt);
* reset tx packet private context (needed to be called between packets) * reset tx packet private context (needed to be called between packets)
* *
* @pkt: packet * @pkt: packet
* @dev: PCI device processing the next packet * @callback: function to free the fragments
* * @context: pointer to be passed to the callback
*/ */
void net_tx_pkt_reset(struct NetTxPkt *pkt, PCIDevice *dev); void net_tx_pkt_reset(struct NetTxPkt *pkt,
NetTxPktFreeFrag callback, void *context);
/** /**
* Unmap a fragment mapped from a PCI device. * Unmap a fragment mapped from a PCI device.
@ -162,6 +161,16 @@ void net_tx_pkt_reset(struct NetTxPkt *pkt, PCIDevice *dev);
*/ */
void net_tx_pkt_unmap_frag_pci(void *context, void *base, size_t len); void net_tx_pkt_unmap_frag_pci(void *context, void *base, size_t len);
/**
* map data fragment from PCI device and populate it into pkt context.
*
* @pci_dev: PCI device owning fragment
* @pa: physical address of fragment
* @len: length of fragment
*/
bool net_tx_pkt_add_raw_fragment_pci(struct NetTxPkt *pkt, PCIDevice *pci_dev,
dma_addr_t pa, size_t len);
/** /**
* Send packet to qemu. handles sw offloads if vhdr is not supported. * Send packet to qemu. handles sw offloads if vhdr is not supported.
* *
@ -182,7 +191,7 @@ bool net_tx_pkt_send(struct NetTxPkt *pkt, NetClientState *nc);
* @ret: operation result * @ret: operation result
*/ */
bool net_tx_pkt_send_custom(struct NetTxPkt *pkt, bool offload, bool net_tx_pkt_send_custom(struct NetTxPkt *pkt, bool offload,
NetTxPktCallback callback, void *context); NetTxPktSend callback, void *context);
/** /**
* parse raw packet data and analyze offload requirements. * parse raw packet data and analyze offload requirements.

View file

@ -651,9 +651,8 @@ static void vmxnet3_process_tx_queue(VMXNET3State *s, int qidx)
data_len = (txd.len > 0) ? txd.len : VMXNET3_MAX_TX_BUF_SIZE; data_len = (txd.len > 0) ? txd.len : VMXNET3_MAX_TX_BUF_SIZE;
data_pa = txd.addr; data_pa = txd.addr;
if (!net_tx_pkt_add_raw_fragment(s->tx_pkt, if (!net_tx_pkt_add_raw_fragment_pci(s->tx_pkt, PCI_DEVICE(s),
data_pa, data_pa, data_len)) {
data_len)) {
s->skip_current_tx_pkt = true; s->skip_current_tx_pkt = true;
} }
} }
@ -678,7 +677,8 @@ static void vmxnet3_process_tx_queue(VMXNET3State *s, int qidx)
vmxnet3_complete_packet(s, qidx, txd_idx); vmxnet3_complete_packet(s, qidx, txd_idx);
s->tx_sop = true; s->tx_sop = true;
s->skip_current_tx_pkt = false; s->skip_current_tx_pkt = false;
net_tx_pkt_reset(s->tx_pkt, PCI_DEVICE(s)); net_tx_pkt_reset(s->tx_pkt,
net_tx_pkt_unmap_frag_pci, PCI_DEVICE(s));
} }
} }
} }
@ -1159,7 +1159,7 @@ static void vmxnet3_deactivate_device(VMXNET3State *s)
{ {
if (s->device_active) { if (s->device_active) {
VMW_CBPRN("Deactivating vmxnet3..."); VMW_CBPRN("Deactivating vmxnet3...");
net_tx_pkt_reset(s->tx_pkt, PCI_DEVICE(s)); net_tx_pkt_reset(s->tx_pkt, net_tx_pkt_unmap_frag_pci, PCI_DEVICE(s));
net_tx_pkt_uninit(s->tx_pkt); net_tx_pkt_uninit(s->tx_pkt);
net_rx_pkt_uninit(s->rx_pkt); net_rx_pkt_uninit(s->rx_pkt);
s->device_active = false; s->device_active = false;
@ -1519,7 +1519,7 @@ static void vmxnet3_activate_device(VMXNET3State *s)
/* Preallocate TX packet wrapper */ /* Preallocate TX packet wrapper */
VMW_CFPRN("Max TX fragments is %u", s->max_tx_frags); VMW_CFPRN("Max TX fragments is %u", s->max_tx_frags);
net_tx_pkt_init(&s->tx_pkt, PCI_DEVICE(s), s->max_tx_frags); net_tx_pkt_init(&s->tx_pkt, s->max_tx_frags);
net_rx_pkt_init(&s->rx_pkt); net_rx_pkt_init(&s->rx_pkt);
/* Read rings memory locations for RX queues */ /* Read rings memory locations for RX queues */
@ -2399,7 +2399,7 @@ static int vmxnet3_post_load(void *opaque, int version_id)
{ {
VMXNET3State *s = opaque; VMXNET3State *s = opaque;
net_tx_pkt_init(&s->tx_pkt, PCI_DEVICE(s), s->max_tx_frags); net_tx_pkt_init(&s->tx_pkt, s->max_tx_frags);
net_rx_pkt_init(&s->rx_pkt); net_rx_pkt_init(&s->rx_pkt);
if (s->msix_used) { if (s->msix_used) {