Home | History | Annotate | Line # | Download | only in libpthread
pthread_lock.c revision 1.29
      1  1.29    skrll /*	$NetBSD: pthread_lock.c,v 1.29 2007/09/24 12:19:39 skrll Exp $	*/
      2   1.2  thorpej 
      3   1.2  thorpej /*-
      4  1.19       ad  * Copyright (c) 2001, 2006, 2007 The NetBSD Foundation, Inc.
      5   1.2  thorpej  * All rights reserved.
      6   1.2  thorpej  *
      7   1.2  thorpej  * This code is derived from software contributed to The NetBSD Foundation
      8  1.19       ad  * by Nathan J. Williams and Andrew Doran.
      9   1.2  thorpej  *
     10   1.2  thorpej  * Redistribution and use in source and binary forms, with or without
     11   1.2  thorpej  * modification, are permitted provided that the following conditions
     12   1.2  thorpej  * are met:
     13   1.2  thorpej  * 1. Redistributions of source code must retain the above copyright
     14   1.2  thorpej  *    notice, this list of conditions and the following disclaimer.
     15   1.2  thorpej  * 2. Redistributions in binary form must reproduce the above copyright
     16   1.2  thorpej  *    notice, this list of conditions and the following disclaimer in the
     17   1.2  thorpej  *    documentation and/or other materials provided with the distribution.
     18   1.2  thorpej  * 3. All advertising materials mentioning features or use of this software
     19   1.2  thorpej  *    must display the following acknowledgement:
     20   1.2  thorpej  *        This product includes software developed by the NetBSD
     21   1.2  thorpej  *        Foundation, Inc. and its contributors.
     22   1.2  thorpej  * 4. Neither the name of The NetBSD Foundation nor the names of its
     23   1.2  thorpej  *    contributors may be used to endorse or promote products derived
     24   1.2  thorpej  *    from this software without specific prior written permission.
     25   1.2  thorpej  *
     26   1.2  thorpej  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     27   1.2  thorpej  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     28   1.2  thorpej  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     29   1.2  thorpej  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     30   1.2  thorpej  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     31   1.2  thorpej  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     32   1.2  thorpej  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     33   1.2  thorpej  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     34   1.2  thorpej  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     35   1.2  thorpej  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     36   1.2  thorpej  * POSSIBILITY OF SUCH DAMAGE.
     37   1.2  thorpej  */
     38   1.6    lukem 
     39  1.23       ad /*
     40  1.23       ad  * libpthread internal spinlock routines.
     41  1.23       ad  */
     42  1.23       ad 
     43   1.6    lukem #include <sys/cdefs.h>
     44  1.29    skrll __RCSID("$NetBSD: pthread_lock.c,v 1.29 2007/09/24 12:19:39 skrll Exp $");
     45   1.2  thorpej 
     46  1.12       he #include <sys/types.h>
     47  1.11       cl #include <sys/lock.h>
     48   1.2  thorpej #include <sys/ras.h>
     49   1.2  thorpej 
     50   1.2  thorpej #include <errno.h>
     51   1.2  thorpej #include <unistd.h>
     52  1.19       ad #include <stdio.h>
     53  1.22       ad #include <stdlib.h>
     54   1.2  thorpej 
     55   1.2  thorpej #include "pthread.h"
     56   1.2  thorpej #include "pthread_int.h"
     57   1.2  thorpej 
     58  1.22       ad /* How many times to try acquiring spin locks on MP systems. */
     59  1.22       ad #define	PTHREAD__NSPINS		1024
     60  1.22       ad 
     61   1.5  nathanw #ifdef PTHREAD_SPIN_DEBUG_PRINT
     62   1.2  thorpej #define SDPRINTF(x) DPRINTF(x)
     63   1.2  thorpej #else
     64   1.2  thorpej #define SDPRINTF(x)
     65   1.2  thorpej #endif
     66   1.2  thorpej 
     67  1.23       ad static void pthread_spinlock_slow(pthread_spin_t *);
     68  1.23       ad 
     69  1.10  thorpej RAS_DECL(pthread__lock);
     70   1.2  thorpej 
     71  1.27    skrll int
     72  1.27    skrll pthread__simple_locked_p(__cpu_simple_lock_t *alp)
     73  1.27    skrll {
     74  1.27    skrll 	return __SIMPLELOCK_LOCKED_P(alp);
     75  1.27    skrll }
     76  1.27    skrll 
     77  1.29    skrll #ifdef PTHREAD__ASM_RASOPS
     78  1.29    skrll 
     79  1.29    skrll void pthread__ras_simple_lock_init(__cpu_simple_lock_t *);
     80  1.29    skrll int pthread__ras_simple_lock_try(__cpu_simple_lock_t *);
     81  1.29    skrll void pthread__ras_simple_unlock(__cpu_simple_lock_t *);
     82  1.29    skrll 
     83  1.29    skrll #else
     84  1.29    skrll 
     85  1.29    skrll static void
     86  1.29    skrll pthread__ras_simple_lock_init(__cpu_simple_lock_t *alp)
     87   1.2  thorpej {
     88   1.2  thorpej 
     89  1.27    skrll 	__cpu_simple_lock_clear(alp);
     90   1.2  thorpej }
     91   1.2  thorpej 
     92  1.29    skrll static int
     93  1.29    skrll pthread__ras_simple_lock_try(__cpu_simple_lock_t *alp)
     94   1.2  thorpej {
     95  1.28    skrll 	int locked;
     96   1.2  thorpej 
     97  1.10  thorpej 	RAS_START(pthread__lock);
     98  1.28    skrll 	locked = __SIMPLELOCK_LOCKED_P(alp);
     99  1.27    skrll 	__cpu_simple_lock_set(alp);
    100  1.10  thorpej 	RAS_END(pthread__lock);
    101   1.2  thorpej 
    102  1.28    skrll 	return !locked;
    103   1.2  thorpej }
    104   1.2  thorpej 
    105  1.29    skrll static void
    106  1.29    skrll pthread__ras_simple_unlock(__cpu_simple_lock_t *alp)
    107  1.29    skrll {
    108  1.29    skrll 
    109  1.29    skrll 	__cpu_simple_lock_clear(alp);
    110  1.29    skrll }
    111  1.29    skrll 
    112  1.29    skrll #endif /* PTHREAD__ASM_RASOPS */
    113  1.29    skrll 
    114  1.29    skrll static const struct pthread_lock_ops pthread__lock_ops_ras = {
    115  1.29    skrll 	pthread__ras_simple_lock_init,
    116  1.29    skrll 	pthread__ras_simple_lock_try,
    117  1.29    skrll 	pthread__ras_simple_unlock,
    118  1.29    skrll };
    119  1.29    skrll 
    120  1.29    skrll static void
    121  1.29    skrll pthread__atomic_simple_lock_init(__cpu_simple_lock_t *alp)
    122  1.29    skrll {
    123  1.29    skrll 
    124  1.29    skrll 	__cpu_simple_lock_init(alp);
    125  1.29    skrll }
    126  1.29    skrll 
    127  1.29    skrll static int
    128  1.29    skrll pthread__atomic_simple_lock_try(__cpu_simple_lock_t *alp)
    129  1.29    skrll {
    130  1.29    skrll 
    131  1.29    skrll 	return (__cpu_simple_lock_try(alp));
    132  1.29    skrll }
    133  1.29    skrll 
    134  1.29    skrll static void
    135  1.29    skrll pthread__atomic_simple_unlock(__cpu_simple_lock_t *alp)
    136   1.2  thorpej {
    137   1.2  thorpej 
    138  1.23       ad 	__cpu_simple_unlock(alp);
    139  1.29    skrll }
    140  1.29    skrll 
    141  1.29    skrll static const struct pthread_lock_ops pthread__lock_ops_atomic = {
    142  1.29    skrll 	pthread__atomic_simple_lock_init,
    143  1.29    skrll 	pthread__atomic_simple_lock_try,
    144  1.29    skrll 	pthread__atomic_simple_unlock,
    145  1.29    skrll };
    146  1.27    skrll 
    147  1.29    skrll /*
    148  1.29    skrll  * We default to pointing to the RAS primitives; we might need to use
    149  1.29    skrll  * locks early, but before main() starts.  This is safe, since no other
    150  1.29    skrll  * threads will be active for the process, so atomicity will not be
    151  1.29    skrll  * required.
    152  1.29    skrll  */
    153  1.29    skrll const struct pthread_lock_ops *pthread__lock_ops = &pthread__lock_ops_ras;
    154   1.2  thorpej 
    155   1.2  thorpej void
    156  1.23       ad pthread_spinlock(pthread_spin_t *lock)
    157   1.2  thorpej {
    158  1.23       ad #ifdef PTHREAD_SPIN_DEBUG
    159  1.23       ad 	pthread_t thread = pthread__self();
    160  1.22       ad 
    161  1.23       ad 	SDPRINTF(("(pthread_spinlock %p) spinlock %p (count %d)\n",
    162  1.23       ad 	    thread, lock, thread->pt_spinlocks));
    163  1.23       ad 	pthread__assert(thread->pt_spinlocks >= 0);
    164  1.23       ad 	thread->pt_spinlocks++;
    165  1.24       ad 	PTHREADD_ADD(PTHREADD_SPINLOCKS);
    166  1.23       ad #endif
    167  1.20       ad 
    168  1.29    skrll 	if (__predict_true(pthread__simple_lock_try(lock)))
    169  1.29    skrll 		return;
    170  1.24       ad 
    171  1.24       ad 	pthread_spinlock_slow(lock);
    172   1.2  thorpej }
    173   1.2  thorpej 
    174  1.23       ad /*
    175  1.23       ad  * Prevent this routine from being inlined.  The common case is no
    176  1.23       ad  * contention and it's better to not burden the instruction decoder.
    177  1.23       ad  */
    178  1.23       ad #if __GNUC_PREREQ__(3, 0)
    179  1.23       ad __attribute ((noinline))
    180  1.23       ad #endif
    181  1.23       ad static void
    182  1.23       ad pthread_spinlock_slow(pthread_spin_t *lock)
    183   1.2  thorpej {
    184  1.21       ad 	int count;
    185   1.2  thorpej #ifdef PTHREAD_SPIN_DEBUG
    186  1.23       ad 	pthread_t thread = pthread__self();
    187   1.2  thorpej #endif
    188  1.19       ad 
    189   1.2  thorpej 	do {
    190  1.21       ad 		count = pthread__nspins;
    191  1.27    skrll 		while (pthread__simple_locked_p(lock) && --count > 0)
    192  1.21       ad 			pthread__smt_pause();
    193  1.21       ad 		if (count > 0) {
    194  1.21       ad 			if (pthread__simple_lock_try(lock))
    195  1.21       ad 				break;
    196  1.21       ad 			continue;
    197  1.17       ad 		}
    198   1.2  thorpej 
    199  1.23       ad #ifdef PTHREAD_SPIN_DEBUG
    200  1.19       ad 		SDPRINTF(("(pthread_spinlock %p) retrying spinlock %p "
    201  1.19       ad 		    "(count %d)\n", thread, lock,
    202  1.19       ad 		    thread->pt_spinlocks));
    203  1.19       ad 		thread->pt_spinlocks--;
    204  1.26       ad 		/* XXXLWP far from ideal */
    205  1.26       ad 		sched_yield();
    206  1.19       ad 		thread->pt_spinlocks++;
    207  1.23       ad #else
    208  1.26       ad 		/* XXXLWP far from ideal */
    209  1.26       ad 		sched_yield();
    210  1.23       ad #endif
    211  1.19       ad 	} while (/*CONSTCOND*/ 1);
    212   1.2  thorpej }
    213   1.2  thorpej 
    214   1.2  thorpej int
    215  1.23       ad pthread_spintrylock(pthread_spin_t *lock)
    216   1.2  thorpej {
    217  1.23       ad #ifdef PTHREAD_SPIN_DEBUG
    218  1.23       ad 	pthread_t thread = pthread__self();
    219   1.2  thorpej 	int ret;
    220   1.2  thorpej 
    221  1.19       ad 	SDPRINTF(("(pthread_spintrylock %p) spinlock %p (count %d)\n",
    222  1.19       ad 	    thread, lock, thread->pt_spinlocks));
    223  1.19       ad 	thread->pt_spinlocks++;
    224   1.2  thorpej 	ret = pthread__simple_lock_try(lock);
    225  1.18       ad 	if (!ret)
    226  1.19       ad 		thread->pt_spinlocks--;
    227   1.2  thorpej 	return ret;
    228  1.23       ad #else
    229  1.23       ad 	return pthread__simple_lock_try(lock);
    230  1.23       ad #endif
    231   1.2  thorpej }
    232   1.2  thorpej 
    233   1.2  thorpej void
    234  1.23       ad pthread_spinunlock(pthread_spin_t *lock)
    235   1.2  thorpej {
    236  1.23       ad #ifdef PTHREAD_SPIN_DEBUG
    237  1.23       ad 	pthread_t thread = pthread__self();
    238   1.2  thorpej 
    239  1.19       ad 	SDPRINTF(("(pthread_spinunlock %p) spinlock %p (count %d)\n",
    240  1.19       ad 	    thread, lock, thread->pt_spinlocks));
    241  1.19       ad 
    242   1.2  thorpej 	pthread__simple_unlock(lock);
    243  1.19       ad 	thread->pt_spinlocks--;
    244   1.5  nathanw 	pthread__assert(thread->pt_spinlocks >= 0);
    245  1.23       ad 	PTHREADD_ADD(PTHREADD_SPINUNLOCKS);
    246  1.23       ad #else
    247  1.23       ad 	pthread__simple_unlock(lock);
    248   1.2  thorpej #endif
    249   1.2  thorpej }
    250   1.2  thorpej 
    251  1.23       ad /*
    252  1.23       ad  * Initialize the locking primitives.  On uniprocessors, we always
    253  1.23       ad  * use Restartable Atomic Sequences if they are available.  Otherwise,
    254  1.23       ad  * we fall back onto machine-dependent atomic lock primitives.
    255   1.2  thorpej  */
    256  1.23       ad void
    257  1.23       ad pthread__lockprim_init(void)
    258   1.2  thorpej {
    259  1.23       ad 	char *p;
    260   1.2  thorpej 
    261  1.23       ad 	if ((p = getenv("PTHREAD_NSPINS")) != NULL)
    262  1.23       ad 		pthread__nspins = atoi(p);
    263  1.23       ad 	else if (pthread__concurrency != 1)
    264  1.23       ad 		pthread__nspins = PTHREAD__NSPINS;
    265  1.23       ad 	else
    266  1.23       ad 		pthread__nspins = 1;
    267  1.19       ad 
    268  1.23       ad 	if (pthread__concurrency != 1) {
    269  1.29    skrll 		pthread__lock_ops = &pthread__lock_ops_atomic;
    270  1.23       ad 		return;
    271  1.23       ad 	}
    272   1.2  thorpej 
    273  1.23       ad 	if (rasctl(RAS_ADDR(pthread__lock), RAS_SIZE(pthread__lock),
    274  1.23       ad 	    RAS_INSTALL) != 0) {
    275  1.29    skrll 		pthread__lock_ops = &pthread__lock_ops_atomic;
    276  1.29    skrll 		return;
    277  1.17       ad 	}
    278   1.2  thorpej }
    279   1.2  thorpej 
    280  1.23       ad void
    281  1.23       ad pthread_lockinit(pthread_spin_t *lock)
    282   1.2  thorpej {
    283   1.2  thorpej 
    284  1.23       ad 	pthread__simple_lock_init(lock);
    285   1.2  thorpej }
    286