Source-Changes-HG archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
[src/netbsd-6]: src/sys/arch/mips Pull up following revision(s) (requested by...
details: https://anonhg.NetBSD.org/src/rev/6bd832c68ac2
branches: netbsd-6
changeset: 776650:6bd832c68ac2
user: msaitoh <msaitoh%NetBSD.org@localhost>
date: Wed Jun 11 15:38:04 2014 +0000
description:
Pull up following revision(s) (requested by skrll in ticket #1068):
sys/arch/mips/mips/pmap.c: revision 1.214
sys/arch/mips/include/pmap.h: revision 1.63
sys/arch/mips/mips/pmap_segtab.c: revision 1.8
Deal with incompatible cache aliases. Specifically,
- always flush an ephemeral page on unmap
- track unmanaged mappings (mappings entered via pmap_kenter_pa) for
aliases where required and handle appropriately (via pmap_enter_pv)
Hopefully this (finally) addresses the instability reported in the
following PRs:
PR/44900 - R5000/Rm5200 mips ports are broken
PR/46890 - upcoming NetBSD 6.0 release is very unstable/unusable on cobalt qube2
PR/48628 - cobalt and hpcmips ports are dead
diffstat:
sys/arch/mips/include/pmap.h | 4 +-
sys/arch/mips/mips/pmap.c | 141 +++++++++++++++++++++-----------------
sys/arch/mips/mips/pmap_segtab.c | 17 +---
3 files changed, 83 insertions(+), 79 deletions(-)
diffs (truncated from 441 to 300 lines):
diff -r 6771a0490065 -r 6bd832c68ac2 sys/arch/mips/include/pmap.h
--- a/sys/arch/mips/include/pmap.h Fri Jun 06 05:10:59 2014 +0000
+++ b/sys/arch/mips/include/pmap.h Wed Jun 11 15:38:04 2014 +0000
@@ -1,4 +1,4 @@
-/* $NetBSD: pmap.h,v 1.61.8.1 2012/07/05 18:39:42 riz Exp $ */
+/* $NetBSD: pmap.h,v 1.61.8.2 2014/06/11 15:38:04 msaitoh Exp $ */
/*
* Copyright (c) 1992, 1993
@@ -283,6 +283,7 @@
#endif /* MIPS3_PLUS */
#define PMAP_STEAL_MEMORY /* enable pmap_steal_memory() */
+#define PMAP_ENABLE_PMAP_KMPAGE /* enable the PMAP_KMPAGE flag */
/*
* Alternate mapping hooks for pool pages. Avoids thrashing the TLB.
@@ -329,6 +330,7 @@
struct pv_entry *pv_next; /* next pv_entry */
struct pmap *pv_pmap; /* pmap where mapping lies */
vaddr_t pv_va; /* virtual address for mapping */
+#define PV_KENTER 0x001
} *pv_entry_t;
#define PG_MD_UNCACHED 0x0001 /* page is mapped uncached */
diff -r 6771a0490065 -r 6bd832c68ac2 sys/arch/mips/mips/pmap.c
--- a/sys/arch/mips/mips/pmap.c Fri Jun 06 05:10:59 2014 +0000
+++ b/sys/arch/mips/mips/pmap.c Wed Jun 11 15:38:04 2014 +0000
@@ -1,4 +1,4 @@
-/* $NetBSD: pmap.c,v 1.207.2.2 2014/05/21 20:39:17 bouyer Exp $ */
+/* $NetBSD: pmap.c,v 1.207.2.3 2014/06/11 15:38:05 msaitoh Exp $ */
/*-
* Copyright (c) 1998, 2001 The NetBSD Foundation, Inc.
@@ -67,7 +67,7 @@
#include <sys/cdefs.h>
-__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.207.2.2 2014/05/21 20:39:17 bouyer Exp $");
+__KERNEL_RCSID(0, "$NetBSD: pmap.c,v 1.207.2.3 2014/06/11 15:38:05 msaitoh Exp $");
/*
* Manages physical address maps.
@@ -317,7 +317,7 @@
/* Forward function declarations */
void pmap_remove_pv(pmap_t, vaddr_t, struct vm_page *, bool);
-void pmap_enter_pv(pmap_t, vaddr_t, struct vm_page *, u_int *);
+void pmap_enter_pv(pmap_t, vaddr_t, struct vm_page *, u_int *, int);
pt_entry_t *pmap_pte(pmap_t, vaddr_t);
/*
@@ -386,13 +386,13 @@
}
PG_MD_PVLIST_UNLOCK(md);
kpreempt_disable();
- pmap_tlb_syncicache(md->pvh_first.pv_va, onproc);
+ pmap_tlb_syncicache(trunc_page(md->pvh_first.pv_va), onproc);
kpreempt_enable();
#else
if (MIPS_HAS_R4K_MMU) {
if (PG_MD_CACHED_P(md)) {
mips_icache_sync_range_index(
- md->pvh_first.pv_va, PAGE_SIZE);
+ trunc_page(md->pvh_first.pv_va), PAGE_SIZE);
}
} else {
mips_icache_sync_range(MIPS_PHYS_TO_KSEG0(VM_PAGE_TO_PHYS(pg)),
@@ -436,10 +436,10 @@
*/
(void)PG_MD_PVLIST_LOCK(md, false);
if (PG_MD_CACHED_P(md)
- && mips_cache_badalias(pv->pv_va, va))
- mips_dcache_wbinv_range_index(pv->pv_va, PAGE_SIZE);
- if (pv->pv_pmap == NULL)
- pv->pv_va = va;
+ && mips_cache_badalias(pv->pv_va, va)) {
+ mips_dcache_wbinv_range_index(trunc_page(pv->pv_va),
+ PAGE_SIZE);
+ }
PG_MD_PVLIST_UNLOCK(md);
}
@@ -450,23 +450,13 @@
pmap_unmap_ephemeral_page(struct vm_page *pg, vaddr_t va,
pt_entry_t old_pt_entry)
{
- struct vm_page_md * const md = VM_PAGE_TO_MD(pg);
- pv_entry_t pv = &md->pvh_first;
-
+
if (MIPS_CACHE_VIRTUAL_ALIAS) {
- (void)PG_MD_PVLIST_LOCK(md, false);
- if (PG_MD_CACHED_P(md)
- || (pv->pv_pmap != NULL
- && mips_cache_badalias(pv->pv_va, va))) {
-
- /*
- * If this page was previously cached or we had to use an
- * incompatible alias and it has a valid mapping, flush it
- * from the cache.
- */
- mips_dcache_wbinv_range(va, PAGE_SIZE);
- }
- PG_MD_PVLIST_UNLOCK(md);
+ /*
+ * Flush the page to avoid future incompatible aliases
+ */
+ KASSERT((va & PAGE_MASK) == 0);
+ mips_dcache_wbinv_range(va, PAGE_SIZE);
}
#ifndef _LP64
/*
@@ -1073,7 +1063,7 @@
while (pv != NULL) {
const pmap_t pmap = pv->pv_pmap;
const uint16_t gen = PG_MD_PVLIST_GEN(md);
- va = pv->pv_va;
+ va = trunc_page(pv->pv_va);
PG_MD_PVLIST_UNLOCK(md);
pmap_protect(pmap, va, va + PAGE_SIZE, prot);
KASSERT(pv->pv_pmap == pmap);
@@ -1101,7 +1091,7 @@
pv = &md->pvh_first;
while (pv->pv_pmap != NULL) {
const pmap_t pmap = pv->pv_pmap;
- va = pv->pv_va;
+ va = trunc_page(pv->pv_va);
PG_MD_PVLIST_UNLOCK(md);
pmap_remove(pmap, va, va + PAGE_SIZE);
pmap_update(pmap);
@@ -1118,6 +1108,9 @@
const uint32_t pg_mask = ~(mips_pg_m_bit() | mips_pg_ro_bit());
const uint32_t p = (flags & VM_PROT_WRITE) ? mips_pg_rw_bit() : mips_pg_ro_bit();
KASSERT(kpreempt_disabled());
+ KASSERT((sva & PAGE_MASK) == 0);
+ KASSERT((eva & PAGE_MASK) == 0);
+
/*
* Change protection on every valid mapping within this segment.
*/
@@ -1162,6 +1155,8 @@
pt_entry_t *pte;
u_int p;
+ KASSERT((sva & PAGE_MASK) == 0);
+ KASSERT((eva & PAGE_MASK) == 0);
PMAP_COUNT(protect);
#ifdef DEBUG
if (pmapdebug & (PDB_FOLLOW|PDB_PROTECT))
@@ -1315,7 +1310,7 @@
pv != NULL;
pv = pv->pv_next) {
pmap_t pmap = pv->pv_pmap;
- vaddr_t va = pv->pv_va;
+ vaddr_t va = trunc_page(pv->pv_va);
pt_entry_t *pte;
uint32_t pt_entry;
@@ -1493,6 +1488,7 @@
struct vm_page_md * const md = VM_PAGE_TO_MD(pg);
PMAP_COUNT(enter_exec_mapping);
if (!PG_MD_EXECPAGE_P(md)) {
+ KASSERT((pa & PAGE_MASK) == 0);
mips_icache_sync_range(MIPS_PHYS_TO_KSEG0(pa),
PAGE_SIZE);
pmap_set_mdpage_attributes(md, PG_MD_EXECPAGE);
@@ -1505,7 +1501,7 @@
kpreempt_disable();
if (pmap == pmap_kernel()) {
if (pg)
- pmap_enter_pv(pmap, va, pg, &npte);
+ pmap_enter_pv(pmap, va, pg, &npte, 0);
/* enter entries into kernel pmap */
pte = kvtopte(va);
@@ -1546,7 +1542,7 @@
/* Done after case that may sleep/return. */
if (pg)
- pmap_enter_pv(pmap, va, pg, &npte);
+ pmap_enter_pv(pmap, va, pg, &npte, 0);
/*
* Now validate mapping with desired protection/wiring.
@@ -1663,6 +1659,17 @@
kpreempt_disable();
pte = kvtopte(va);
KASSERT(!mips_pg_v(pte->pt_entry));
+
+ /*
+ * No need to track non-managed pages or PMAP_KMPAGEs pages for aliases
+ */
+ if (managed && (flags & PMAP_KMPAGE) == 0) {
+ pmap_t pmap = pmap_kernel();
+ struct vm_page *pg = PHYS_TO_VM_PAGE(pa);
+
+ pmap_enter_pv(pmap, va, pg, &npte, PV_KENTER);
+ }
+
pte->pt_entry = npte;
pmap_tlb_update_addr(pmap_kernel(), va, npte, false);
kpreempt_enable();
@@ -1688,23 +1695,10 @@
}
PMAP_COUNT(kremove_pages);
- if (MIPS_HAS_R4K_MMU && MIPS_CACHE_VIRTUAL_ALIAS) {
- struct vm_page * const pg =
- PHYS_TO_VM_PAGE(mips_tlbpfn_to_paddr(pt_entry));
- if (pg != NULL) {
- struct vm_page_md * const md = VM_PAGE_TO_MD(pg);
- (void)PG_MD_PVLIST_LOCK(md, false);
- pv_entry_t pv = &md->pvh_first;
- if (pv->pv_pmap == NULL) {
- pv->pv_va = va;
- } else if (PG_MD_CACHED_P(md)
- && mips_cache_badalias(pv->pv_va, va)) {
- mips_dcache_wbinv_range(va, PAGE_SIZE);
- }
- PG_MD_PVLIST_UNLOCK(md);
- }
- }
-
+ struct vm_page * const pg =
+ PHYS_TO_VM_PAGE(mips_tlbpfn_to_paddr(pt_entry));
+ if (pg)
+ pmap_remove_pv(pmap_kernel(), va, pg, false);
pte->pt_entry = new_pt_entry;
pmap_tlb_invalidate_addr(pmap_kernel(), va);
}
@@ -2001,10 +1995,13 @@
gen = PG_MD_PVLIST_LOCK(md, false);
for (; pv != NULL; pv = pv_next) {
pmap_t pmap = pv->pv_pmap;
- vaddr_t va = pv->pv_va;
+ vaddr_t va = trunc_page(pv->pv_va);
pt_entry_t *pte;
uint32_t pt_entry;
+
pv_next = pv->pv_next;
+ if (pv->pv_va & PV_KENTER)
+ continue;
if (pmap == pmap_kernel()) {
pte = kvtopte(va);
} else {
@@ -2083,8 +2080,13 @@
#ifdef _LP64
KASSERT(!MIPS_XKPHYS_P(pv->pv_va));
#endif
+ pv_entry_t opv = &md->pvh_first;
+ for (; opv != NULL; opv = opv->pv_next) {
+ if (mips_cache_badalias(pv->pv_va, opv->pv_va)) {
+ KASSERT(PG_MD_UNCACHED_P(md));
+ }
+ }
}
- pv = &md->pvh_first;
}
#endif /* PARANOIADIAG */
}
@@ -2094,7 +2096,8 @@
* physical to virtual map table.
*/
void
-pmap_enter_pv(pmap_t pmap, vaddr_t va, struct vm_page *pg, u_int *npte)
+pmap_enter_pv(pmap_t pmap, vaddr_t va, struct vm_page *pg, u_int *npte,
+ int flags)
{
struct vm_page_md * const md = VM_PAGE_TO_MD(pg);
pv_entry_t pv, npv, apv;
@@ -2133,7 +2136,7 @@
PMAP_COUNT(mappings);
pmap_clear_mdpage_attributes(md, PG_MD_UNCACHED);
pv->pv_pmap = pmap;
- pv->pv_va = va;
+ pv->pv_va = va | flags;
} else {
#if defined(MIPS3_PLUS) && !defined(MULTIPROCESSOR) /* XXX mmu XXX */
if (MIPS_CACHE_VIRTUAL_ALIAS) {
@@ -2154,8 +2157,9 @@
if (mips_cache_badalias(pv->pv_va, va)) {
for (npv = pv; npv; npv = npv->pv_next) {
- pmap_remove(npv->pv_pmap, npv->pv_va,
- npv->pv_va + PAGE_SIZE);
+ vaddr_t nva = trunc_page(npv->pv_va);
+ pmap_remove(npv->pv_pmap, nva,
+ nva + PAGE_SIZE);
pmap_update(npv->pv_pmap);
goto again;
}
@@ -2174,9 +2178,10 @@
* share the same cache index again.
*/
if (mips_cache_badalias(pv->pv_va, va)) {
+ vaddr_t nva = trunc_page(pv->pv_va);
pmap_page_cache(pg, false);
- mips_dcache_wbinv_range_index(
- pv->pv_va, PAGE_SIZE);
Home |
Main Index |
Thread Index |
Old Index