Source-Changes-HG archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index][Old Index]
[src/trunk]: src/sys/arch Part III of ad's performance improvements for aarch64
details: https://anonhg.NetBSD.org/src/rev/fe9505f2caf8
branches: trunk
changeset: 974817:fe9505f2caf8
user: skrll <skrll%NetBSD.org@localhost>
date: Wed Aug 12 13:28:46 2020 +0000
description:
Part III of ad's performance improvements for aarch64
- Assembly language stubs for mutex_enter() and mutex_exit().
diffstat:
sys/arch/aarch64/aarch64/lock_stubs.S | 81 +++++++++++++++++++++++++++++++++++
sys/arch/aarch64/conf/files.aarch64 | 3 +-
sys/arch/aarch64/include/mutex.h | 4 +-
sys/arch/evbarm/include/mutex.h | 6 ++-
4 files changed, 91 insertions(+), 3 deletions(-)
diffs (124 lines):
diff -r f4f2f244a794 -r fe9505f2caf8 sys/arch/aarch64/aarch64/lock_stubs.S
--- /dev/null Thu Jan 01 00:00:00 1970 +0000
+++ b/sys/arch/aarch64/aarch64/lock_stubs.S Wed Aug 12 13:28:46 2020 +0000
@@ -0,0 +1,81 @@
+/* $NetBSD: lock_stubs.S,v 1.1 2020/08/12 13:28:46 skrll Exp $ */
+
+/*-
+ * Copyright (c) 2014, 2020 The NetBSD Foundation, Inc.
+ * All rights reserved.
+ *
+ * This code is derived from software contributed to The NetBSD Foundation
+ * by Matt Thomas of 3am Software Foundry, and by Andrew Doran.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ * 1. Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in the
+ * documentation and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
+ * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
+ * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
+ * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
+ * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#include "opt_lockdebug.h"
+
+#include <aarch64/asm.h>
+
+#include "assym.h"
+
+RCSID("$NetBSD: lock_stubs.S,v 1.1 2020/08/12 13:28:46 skrll Exp $")
+
+#ifndef LOCKDEBUG
+/*
+ * mutex_enter(): the compare-and-set must be atomic with respect to
+ * interrupts and with respect to other CPUs.
+ */
+ENTRY(mutex_enter)
+ mrs x1, tpidr_el1 /* x1 = curlwp */
+1:
+ ldxr x2, [x0] /* load old value */
+ cbnz x2, 3f /* equals zero? */
+ stxr w3, x1, [x0] /* store curlwp as new value */
+ cbnz w3, 2f /* succeed? nope, try again. */
+ dmb sy /* membar_enter() */
+ ret
+2:
+ b 1b
+3:
+ b _C_LABEL(mutex_vector_enter)
+END(mutex_enter)
+
+/*
+ * mutex_exit(): the compare-and-set need only be atomic with respect
+ * to interrupts. the cheapest way to achieve that may be to use a
+ * restartable sequence, but the code do that would be quite involved,
+ * so just use ldxr+stxr to achieve the same.
+ */
+ENTRY(mutex_exit)
+ dmb sy /* membar_exit() */
+ mrs x1, tpidr_el1 /* x1 = curlwp */
+1:
+ ldxr x2, [x0] /* load old value */
+ cmp x1, x2 /* equals curlwp? */
+ b.ne 3f /* slow path if different */
+ stxr w3, xzr, [x0] /* store zero as new value */
+ cbnz w3, 2f /* succeed? nope, try again. */
+ ret
+2:
+ b 1b
+3:
+ b _C_LABEL(mutex_vector_exit)
+END(mutex_exit)
+#endif /* !LOCKDEBUG */
diff -r f4f2f244a794 -r fe9505f2caf8 sys/arch/aarch64/conf/files.aarch64
--- a/sys/arch/aarch64/conf/files.aarch64 Wed Aug 12 13:19:35 2020 +0000
+++ b/sys/arch/aarch64/conf/files.aarch64 Wed Aug 12 13:28:46 2020 +0000
@@ -1,4 +1,4 @@
-# $NetBSD: files.aarch64,v 1.26 2020/07/25 22:51:57 riastradh Exp $
+# $NetBSD: files.aarch64,v 1.27 2020/08/12 13:28:46 skrll Exp $
defflag opt_cpuoptions.h AARCH64_ALIGNMENT_CHECK
defflag opt_cpuoptions.h AARCH64_EL0_STACK_ALIGNMENT_CHECK
@@ -100,6 +100,7 @@
file arch/aarch64/aarch64/fusu.S
file arch/aarch64/aarch64/idle_machdep.S
file arch/aarch64/aarch64/kobj_machdep.c modular
+file arch/aarch64/aarch64/lock_stubs.S
file arch/aarch64/aarch64/process_machdep.c
file arch/aarch64/aarch64/procfs_machdep.c procfs
file arch/aarch64/aarch64/sig_machdep.c
diff -r f4f2f244a794 -r fe9505f2caf8 sys/arch/aarch64/include/mutex.h
--- a/sys/arch/aarch64/include/mutex.h Wed Aug 12 13:19:35 2020 +0000
+++ b/sys/arch/aarch64/include/mutex.h Wed Aug 12 13:28:46 2020 +0000
@@ -1,3 +1,5 @@
-/* $NetBSD: mutex.h,v 1.1 2014/08/10 05:47:38 matt Exp $ */
+/* $NetBSD: mutex.h,v 1.2 2020/08/12 13:28:46 skrll Exp $ */
#include <arm/mutex.h>
+
+#define __HAVE_MUTEX_STUBS 1
diff -r f4f2f244a794 -r fe9505f2caf8 sys/arch/evbarm/include/mutex.h
--- a/sys/arch/evbarm/include/mutex.h Wed Aug 12 13:19:35 2020 +0000
+++ b/sys/arch/evbarm/include/mutex.h Wed Aug 12 13:28:46 2020 +0000
@@ -1,3 +1,7 @@
-/* $NetBSD: mutex.h,v 1.2 2007/02/09 21:55:03 ad Exp $ */
+/* $NetBSD: mutex.h,v 1.3 2020/08/12 13:28:46 skrll Exp $ */
+#ifdef __aarch64__
+#include <aarch64/mutex.h>
+#else
#include <arm/mutex.h>
+#endif
Home |
Main Index |
Thread Index |
Old Index