Source-Changes-HG archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
[src/trunk]: src/sys/arch/xen/xen convert to bus_dma(9), remove now not neces...
details: https://anonhg.NetBSD.org/src/rev/dfe78ca56ef6
branches: trunk
changeset: 930626:dfe78ca56ef6
user: jdolecek <jdolecek%NetBSD.org@localhost>
date: Fri Apr 10 18:03:06 2020 +0000
description:
convert to bus_dma(9), remove now not necessary XENPVHVM redefines
diffstat:
sys/arch/xen/xen/if_xennet_xenbus.c | 245 ++++++++++++++++-------------------
1 files changed, 115 insertions(+), 130 deletions(-)
diffs (truncated from 443 to 300 lines):
diff -r e033300a62e1 -r dfe78ca56ef6 sys/arch/xen/xen/if_xennet_xenbus.c
--- a/sys/arch/xen/xen/if_xennet_xenbus.c Fri Apr 10 17:26:46 2020 +0000
+++ b/sys/arch/xen/xen/if_xennet_xenbus.c Fri Apr 10 18:03:06 2020 +0000
@@ -1,4 +1,4 @@
-/* $NetBSD: if_xennet_xenbus.c,v 1.109 2020/04/07 11:47:06 jdolecek Exp $ */
+/* $NetBSD: if_xennet_xenbus.c,v 1.110 2020/04/10 18:03:06 jdolecek Exp $ */
/*
* Copyright (c) 2006 Manuel Bouyer.
@@ -81,7 +81,7 @@
*/
#include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: if_xennet_xenbus.c,v 1.109 2020/04/07 11:47:06 jdolecek Exp $");
+__KERNEL_RCSID(0, "$NetBSD: if_xennet_xenbus.c,v 1.110 2020/04/10 18:03:06 jdolecek Exp $");
#include "opt_xen.h"
#include "opt_nfs_boot.h"
@@ -143,21 +143,6 @@
#define DPRINTFN(n,x)
#endif
-#ifdef XENPVHVM
-/* Glue for p2m table stuff. Should be removed eventually */
-#define xpmap_mtop_masked(mpa) (mpa & ~PAGE_MASK)
-#define xpmap_mtop(mpa) (mpa & ~PTE_4KFRAME)
-#define xpmap_ptom_masked(mpa) (mpa & ~PAGE_MASK)
-#define xpmap_ptom(mpa) (mpa & ~PTE_4KFRAME)
-#define xpmap_ptom_map(ppa, mpa)
-#define xpmap_ptom_unmap(ppa)
-#define xpmap_ptom_isvalid 1 /* XXX: valid PA check */
-#define xpmap_pg_nx pmap_pg_nx /* We use the native setting */
-#define xpq_flush_queue() tlbflush()
-#endif /* XENPVHVM */
-
-extern pt_entry_t xpmap_pg_nx;
-
#define GRANT_INVALID_REF -1 /* entry is free */
#define NET_TX_RING_SIZE __CONST_RING_SIZE(netif_tx, PAGE_SIZE)
@@ -168,15 +153,15 @@
uint16_t txreq_id; /* ID passed to backend */
grant_ref_t txreq_gntref; /* grant ref of this request */
struct mbuf *txreq_m; /* mbuf being transmitted */
+ bus_dmamap_t txreq_dmamap;
};
struct xennet_rxreq {
SLIST_ENTRY(xennet_rxreq) rxreq_next;
uint16_t rxreq_id; /* ID passed to backend */
grant_ref_t rxreq_gntref; /* grant ref of this request */
-/* va/pa for this receive buf. ma will be provided by backend */
- paddr_t rxreq_pa;
- vaddr_t rxreq_va;
+ struct mbuf *rxreq_m;
+ bus_dmamap_t rxreq_dmamap;
};
struct xennet_xenbus_softc {
@@ -184,6 +169,7 @@
struct ethercom sc_ethercom;
uint8_t sc_enaddr[6];
struct xenbus_device *sc_xbusd;
+ bus_dma_tag_t sc_dmat;
netif_tx_front_ring_t sc_tx_ring;
netif_rx_front_ring_t sc_rx_ring;
@@ -272,17 +258,13 @@
unsigned long uval;
extern int ifqmaxlen; /* XXX */
char mac[32];
-#ifdef XENNET_DEBUG
- char **dir;
- int dir_n = 0;
- char id_str[20];
-#endif
aprint_normal(": Xen Virtual Network Interface\n");
sc->sc_dev = self;
sc->sc_xbusd = xa->xa_xbusd;
sc->sc_xbusd->xbusd_otherend_changed = xennet_backend_changed;
+ sc->sc_dmat = xa->xa_dmat;
/* xenbus ensure 2 devices can't be probed at the same time */
if (if_xennetrxbuf_cache_inited == 0) {
@@ -295,7 +277,14 @@
mutex_init(&sc->sc_tx_lock, MUTEX_DEFAULT, IPL_NET);
SLIST_INIT(&sc->sc_txreq_head);
for (i = 0; i < NET_TX_RING_SIZE; i++) {
- sc->sc_txreqs[i].txreq_id = i;
+ struct xennet_txreq *txreq = &sc->sc_txreqs[i];
+
+ txreq->txreq_id = i;
+ if (bus_dmamap_create(sc->sc_dmat, PAGE_SIZE, 1, PAGE_SIZE,
+ PAGE_SIZE, BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW,
+ &txreq->txreq_dmamap) != 0)
+ break;
+
SLIST_INSERT_HEAD(&sc->sc_txreq_head, &sc->sc_txreqs[i],
txreq_next);
}
@@ -305,9 +294,9 @@
for (i = 0; i < NET_RX_RING_SIZE; i++) {
struct xennet_rxreq *rxreq = &sc->sc_rxreqs[i];
rxreq->rxreq_id = i;
- rxreq->rxreq_va = (vaddr_t)pool_cache_get_paddr(
- if_xennetrxbuf_cache, PR_WAITOK, &rxreq->rxreq_pa);
- if (rxreq->rxreq_va == 0)
+ if (bus_dmamap_create(sc->sc_dmat, PAGE_SIZE, 1, PAGE_SIZE,
+ PAGE_SIZE, BUS_DMA_WAITOK | BUS_DMA_ALLOCNOW,
+ &rxreq->rxreq_dmamap) != 0)
break;
rxreq->rxreq_gntref = GRANT_INVALID_REF;
SLIST_INSERT_HEAD(&sc->sc_rxreq_head, rxreq, rxreq_next);
@@ -443,10 +432,9 @@
xennet_free_rx_buffer(sc);
for (i = 0; i < NET_RX_RING_SIZE; i++) {
struct xennet_rxreq *rxreq = &sc->sc_rxreqs[i];
- if (rxreq->rxreq_va != 0) {
- pool_cache_put_paddr(if_xennetrxbuf_cache,
- (void *)rxreq->rxreq_va, rxreq->rxreq_pa);
- rxreq->rxreq_va = 0;
+ if (rxreq->rxreq_m != NULL) {
+ m_freem(rxreq->rxreq_m);
+ rxreq->rxreq_m = NULL;
}
}
mutex_exit(&sc->sc_rx_lock);
@@ -692,7 +680,6 @@
* in the ring. This allows the backend to use them to communicate with
* frontend when some data is destined to frontend
*/
-
static void
xennet_alloc_rx_buffer(struct xennet_xenbus_softc *sc)
{
@@ -700,6 +687,10 @@
RING_IDX i;
struct xennet_rxreq *req;
int otherend_id, notify;
+ struct mbuf *m;
+ vaddr_t va;
+ paddr_t pa, ma;
+ struct ifnet *ifp = &sc->sc_ethercom.ec_if;
KASSERT(mutex_owned(&sc->sc_rx_lock));
@@ -709,15 +700,54 @@
req = SLIST_FIRST(&sc->sc_rxreq_head);
KASSERT(req != NULL);
KASSERT(req == &sc->sc_rxreqs[req->rxreq_id]);
+ KASSERT(req->rxreq_m == NULL);
+ KASSERT(req->rxreq_gntref = GRANT_INVALID_REF);
+
+ MGETHDR(m, M_DONTWAIT, MT_DATA);
+ if (__predict_false(m == NULL)) {
+ printf("%s: rx no mbuf\n", ifp->if_xname);
+ break;
+ }
+
+ va = (vaddr_t)pool_cache_get_paddr(
+ if_xennetrxbuf_cache, PR_NOWAIT, &pa);
+ if (__predict_false(va == 0)) {
+ printf("%s: rx no cluster\n", ifp->if_xname);
+ m_freem(m);
+ break;
+ }
+
+ MEXTADD(m, va, PAGE_SIZE,
+ M_DEVBUF, xennet_rx_mbuf_free, NULL);
+ m->m_len = m->m_pkthdr.len = PAGE_SIZE;
+ m->m_ext.ext_paddr = pa;
+ m->m_flags |= M_EXT_RW; /* we own the buffer */
+
+ /* Set M_EXT_CLUSTER so that load_mbuf uses m_ext.ext_paddr */
+ m->m_flags |= M_EXT_CLUSTER;
+ if (__predict_false(bus_dmamap_load_mbuf(sc->sc_dmat,
+ req->rxreq_dmamap, m, BUS_DMA_NOWAIT) != 0)) {
+ printf("%s: rx mbuf load failed", ifp->if_xname);
+ m->m_flags &= ~M_EXT_CLUSTER;
+ m_freem(m);
+ break;
+ }
+ m->m_flags &= ~M_EXT_CLUSTER;
+
+ KASSERT(req->rxreq_dmamap->dm_nsegs == 1);
+ ma = req->rxreq_dmamap->dm_segs[0].ds_addr;
+
+ if (xengnt_grant_access(otherend_id, trunc_page(ma),
+ 0, &req->rxreq_gntref) != 0) {
+ m_freem(m);
+ break;
+ }
+
+ req->rxreq_m = m;
+
RING_GET_REQUEST(&sc->sc_rx_ring, req_prod + i)->id =
req->rxreq_id;
- if (xengnt_grant_access(otherend_id,
- xpmap_ptom_masked(req->rxreq_pa),
- 0, &req->rxreq_gntref) != 0) {
- goto out_loop;
- }
-
RING_GET_REQUEST(&sc->sc_rx_ring, req_prod + i)->gref =
req->rxreq_gntref;
@@ -725,16 +755,13 @@
sc->sc_free_rxreql--;
}
-out_loop:
- if (i == 0) {
- return;
+ /* Notify backend if more Rx is possible */
+ if (i > 0) {
+ sc->sc_rx_ring.req_prod_pvt = req_prod + i;
+ RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&sc->sc_rx_ring, notify);
+ if (notify)
+ hypervisor_notify_via_evtchn(sc->sc_evtchn);
}
-
- sc->sc_rx_ring.req_prod_pvt = req_prod + i;
- RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&sc->sc_rx_ring, notify);
- if (notify)
- hypervisor_notify_via_evtchn(sc->sc_evtchn);
- return;
}
/*
@@ -765,6 +792,10 @@
rxreq->rxreq_gntref = GRANT_INVALID_REF;
}
+ if (rxreq->rxreq_m != NULL) {
+ m_freem(rxreq->rxreq_m);
+ rxreq->rxreq_m = NULL;
+ }
}
DPRINTF(("%s: xennet_free_rx_buffer done\n", device_xname(sc->sc_dev)));
}
@@ -827,6 +858,8 @@
KASSERT(req->txreq_id ==
RING_GET_RESPONSE(&sc->sc_tx_ring, i)->id);
KASSERT(xengnt_status(req->txreq_gntref) == 0);
+ KASSERT(req->txreq_m != NULL);
+
if (__predict_false(
RING_GET_RESPONSE(&sc->sc_tx_ring, i)->status !=
NETIF_RSP_OKAY))
@@ -834,7 +867,9 @@
else
if_statinc(ifp, if_opackets);
xengnt_revoke_access(req->txreq_gntref);
+ bus_dmamap_unload(sc->sc_dmat, req->txreq_dmamap);
m_freem(req->txreq_m);
+ req->txreq_m = NULL;
SLIST_INSERT_HEAD(&sc->sc_txreq_head, req, txreq_next);
}
@@ -860,10 +895,7 @@
struct ifnet *ifp = &sc->sc_ethercom.ec_if;
RING_IDX resp_prod, i;
struct xennet_rxreq *req;
- paddr_t pa;
- vaddr_t va;
struct mbuf *m;
- void *pktp;
int more_to_do;
if (sc->sc_backend_status != BEST_CONNECTED)
@@ -891,47 +923,16 @@
xengnt_revoke_access(req->rxreq_gntref);
req->rxreq_gntref = GRANT_INVALID_REF;
- pa = req->rxreq_pa;
- va = req->rxreq_va;
+ m = req->rxreq_m;
+ req->rxreq_m = NULL;
- pktp = (void *)(va + rx->offset);
-#ifdef XENNET_DEBUG_DUMP
- xennet_hex_dump(pktp, rx->status, "r", rx->id);
-#endif
- MGETHDR(m, M_DONTWAIT, MT_DATA);
- if (__predict_false(m == NULL)) {
- printf("%s: rx no mbuf\n", ifp->if_xname);
- if_statinc(ifp, if_ierrors);
- xennet_rx_free_req(sc, req);
- continue;
- }
+ m->m_len = m->m_pkthdr.len = rx->status;
+ bus_dmamap_sync(sc->sc_dmat, req->rxreq_dmamap, 0,
+ m->m_pkthdr.len, BUS_DMASYNC_PREREAD);
+
MCLAIM(m, &sc->sc_ethercom.ec_rx_mowner);
-
m_set_rcvif(m, ifp);
Home |
Main Index |
Thread Index |
Old Index