Source-Changes-HG archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
[src/trunk]: src/sys/arch/mac68k/mac68k Cleanup: GC unused externs, and some...
details: https://anonhg.NetBSD.org/src/rev/f94aff6bac73
branches: trunk
changeset: 471664:f94aff6bac73
user: scottr <scottr%NetBSD.org@localhost>
date: Wed Apr 07 05:59:14 1999 +0000
description:
Cleanup: GC unused externs, and some KNF.
diffstat:
sys/arch/mac68k/mac68k/pmap.c | 79 +++++++++++++++++++-----------------------
1 files changed, 36 insertions(+), 43 deletions(-)
diffs (192 lines):
diff -r 7fffcd803209 -r f94aff6bac73 sys/arch/mac68k/mac68k/pmap.c
--- a/sys/arch/mac68k/mac68k/pmap.c Wed Apr 07 05:47:37 1999 +0000
+++ b/sys/arch/mac68k/mac68k/pmap.c Wed Apr 07 05:59:14 1999 +0000
@@ -1,4 +1,4 @@
-/* $NetBSD: pmap.c,v 1.53 1999/04/05 06:34:01 scottr Exp $ */
+/* $NetBSD: pmap.c,v 1.54 1999/04/07 05:59:14 scottr Exp $ */
/*
* Copyright (c) 1991, 1993
@@ -59,7 +59,7 @@
* Bitwise and/or operations are significantly faster than bitfield
* references so we use them when accessing STE/PTEs in the pmap_pte_*
* macros. Note also that the two are not always equivalent; e.g.:
- * (*pte & PG_PROT)[4] != pte->pg_prot [1]
+ * (*pte & PG_PROT)[4] != pte->pg_prot[1]
* and a couple of routines that deal with protection and wiring take
* some shortcuts that assume the and/or definitions.
*
@@ -135,30 +135,28 @@
#if defined(M68040)
int dowriteback = 1; /* 68040: enable writeback caching */
int dokwriteback = 1; /* 68040: enable writeback caching of kernel AS */
-
-extern vaddr_t pager_sva, pager_eva;
#endif
#else /* ! DEBUG */
#define PMAP_DPRINTF(l, x) /* nothing */
-#endif
+#endif /* DEBUG */
/*
* Get STEs and PTEs for user/kernel address space
*/
#if defined(M68040)
-#define pmap_ste1(m, v) \
+#define pmap_ste1(m, v) \
(&((m)->pm_stab[(vaddr_t)(v) >> SG4_SHIFT1]))
/* XXX assumes physically contiguous ST pages (if more than one) */
#define pmap_ste2(m, v) \
(&((m)->pm_stab[(st_entry_t *)(*(u_int *)pmap_ste1(m, v) & SG4_ADDR1) \
- (m)->pm_stpa + (((v) & SG4_MASK2) >> SG4_SHIFT2)]))
-#define pmap_ste(m, v) \
+#define pmap_ste(m, v) \
(&((m)->pm_stab[(vaddr_t)(v) \
>> (mmutype == MMU_68040 ? SG4_SHIFT1 : SG_ISHIFT)]))
#define pmap_ste_v(m, v) \
- (mmutype == MMU_68040 \
- ? ((*pmap_ste1(m, v) & SG_V) && \
- (*pmap_ste2(m, v) & SG_V)) \
+ (mmutype == MMU_68040 \
+ ? ((*pmap_ste1(m, v) & SG_V) && \
+ (*pmap_ste2(m, v) & SG_V)) \
: (*pmap_ste(m, v) & SG_V))
#else
#define pmap_ste(m, v) (&((m)->pm_stab[(vaddr_t)(v) >> SG_ISHIFT]))
@@ -174,15 +172,13 @@
#define pmap_pte_prot(pte) (*(pte) & PG_PROT)
#define pmap_pte_v(pte) (*(pte) & PG_V)
-#define pmap_pte_set_w(pte, v) \
+#define pmap_pte_set_w(pte, v) \
if (v) *(pte) |= PG_W; else *(pte) &= ~PG_W
#define pmap_pte_set_prot(pte, v) \
if (v) *(pte) |= PG_PROT; else *(pte) &= ~PG_PROT
#define pmap_pte_w_chg(pte, nw) ((nw) ^ pmap_pte_w(pte))
#define pmap_pte_prot_chg(pte, np) ((np) ^ pmap_pte_prot(pte))
-int pmap_page_index(paddr_t pa);
-
/*
* Given a map and a machine independent protection code,
* convert to a m68k protection code.
@@ -213,19 +209,15 @@
st_entry_t *Sysseg;
pt_entry_t *Sysmap, *Sysptmap;
st_entry_t *Segtabzero, *Segtabzeropa;
-vm_size_t Sysptsize = VM_KERNEL_PT_PAGES;
+vsize_t Sysptsize = VM_KERNEL_PT_PAGES;
struct pmap kernel_pmap_store;
vm_map_t st_map, pt_map;
struct vm_map st_map_store, pt_map_store;
-paddr_t avail_start; /* PA of first available physical page */
-paddr_t avail_next; /* Next available physical page */
-int avail_remaining;/* Number of physical free pages left */
-int avail_range; /* Range avail_next is in */
-paddr_t avail_end; /* Set for ps and friends as */
- /* avail_start + avail_remaining. */
-vm_size_t mem_size; /* memory size in bytes */
+paddr_t avail_start; /* PA of first available physical page */
+paddr_t avail_end; /* PA of last available physical page */
+vsize_t mem_size; /* memory size in bytes */
vaddr_t virtual_avail; /* VA of first avail page (after kernel bss)*/
vaddr_t virtual_end; /* VA of last avail page (end of kernel AS) */
int page_cnt; /* number of pages managed by VM system */
@@ -237,9 +229,6 @@
int pv_nfree;
/* The following four variables are defined in pmap_bootstrap.c */
-extern int numranges;
-extern unsigned long low[8];
-extern unsigned long high[8];
extern int vidlen;
#define VIDMAPSIZE btoc(vidlen)
@@ -257,8 +246,6 @@
struct pv_entry *pmap_alloc_pv __P((void));
void pmap_free_pv __P((struct pv_entry *));
void pmap_collect_pv __P((void));
-void pmap_pinit __P((pmap_t));
-void pmap_release __P((pmap_t));
#define PAGE_IS_MANAGED(pa) (pmap_initialized && \
vm_physseg_find(atop((pa)), NULL) != -1)
@@ -287,6 +274,8 @@
void pmap_changebit __P((paddr_t, int, int));
void pmap_enter_ptpage __P((pmap_t, vaddr_t));
void pmap_collect1 __P((pmap_t, paddr_t, vaddr_t));
+void pmap_pinit __P((pmap_t));
+void pmap_release __P((pmap_t));
#ifdef DEBUG
void pmap_pvdump __P((paddr_t));
@@ -318,35 +307,40 @@
}
/*
- * Routine: pmap_procwr
+ * pmap_procwr: [ INTERFACE ]
+ *
+ * Synchronize caches corresponding to [addr, addr+len) in p.
*
- * Function:
- * Synchronize caches corresponding to [addr, addr+len) in
- * p.
- *
- */
+ * Note: no locking is necessary in this function.
+ */
void
pmap_procwr(p, va, len)
- struct proc *p;
- vm_offset_t va;
- u_long len;
+ struct proc *p;
+ vaddr_t va;
+ u_long len;
{
+
(void)cachectl1(0x80000004, va, len, p);
}
/*
- * Initialize the pmap module.
- * Called by vm_init, to initialize any structures that the pmap
- * system needs to map virtual memory.
+ * pmap_init: [ INTERFACE ]
+ *
+ * Initialize the pmap module. Called by vm_init(), to initialize any
+ * structures that the pmap system needs to map virtual memory.
+ *
+ * Note: no locking is necessary in this function.
*/
void
pmap_init()
{
vaddr_t addr, addr2;
- vm_size_t s;
- struct pv_entry *pv;
+ vsize_t s;
+ struct pv_entry *pv;
char *attr;
- int rv, npages, bank;
+ int rv;
+ int npages;
+ int bank;
PMAP_DPRINTF(PDB_FOLLOW, ("pmap_init()\n"));
@@ -1098,10 +1092,9 @@
/*
* For user mapping, allocate kernel VM resources if necessary.
*/
- if (pmap->pm_ptab == NULL) {
+ if (pmap->pm_ptab == NULL)
pmap->pm_ptab = (pt_entry_t *)
uvm_km_valloc_wait(pt_map, MAC_MAX_PTSIZE);
- }
/*
* Segment table entry not valid, we need a new PT page
Home |
Main Index |
Thread Index |
Old Index