Home | History | Annotate | Line # | Download | only in include
mutex.h revision 1.4.4.2
      1  1.4.4.2  yamt /*	$NetBSD: mutex.h,v 1.4.4.2 2007/02/26 09:06:43 yamt Exp $	*/
      2  1.4.4.2  yamt 
      3  1.4.4.2  yamt /*-
      4  1.4.4.2  yamt  * Copyright (c) 2002, 2007 The NetBSD Foundation, Inc.
      5  1.4.4.2  yamt  * All rights reserved.
      6  1.4.4.2  yamt  *
      7  1.4.4.2  yamt  * This code is derived from software contributed to The NetBSD Foundation
      8  1.4.4.2  yamt  * by Jason R. Thorpe and Andrew Doran.
      9  1.4.4.2  yamt  *
     10  1.4.4.2  yamt  * Redistribution and use in source and binary forms, with or without
     11  1.4.4.2  yamt  * modification, are permitted provided that the following conditions
     12  1.4.4.2  yamt  * are met:
     13  1.4.4.2  yamt  * 1. Redistributions of source code must retain the above copyright
     14  1.4.4.2  yamt  *    notice, this list of conditions and the following disclaimer.
     15  1.4.4.2  yamt  * 2. Redistributions in binary form must reproduce the above copyright
     16  1.4.4.2  yamt  *    notice, this list of conditions and the following disclaimer in the
     17  1.4.4.2  yamt  *    documentation and/or other materials provided with the distribution.
     18  1.4.4.2  yamt  * 3. All advertising materials mentioning features or use of this software
     19  1.4.4.2  yamt  *    must display the following acknowledgement:
     20  1.4.4.2  yamt  *	This product includes software developed by the NetBSD
     21  1.4.4.2  yamt  *	Foundation, Inc. and its contributors.
     22  1.4.4.2  yamt  * 4. Neither the name of The NetBSD Foundation nor the names of its
     23  1.4.4.2  yamt  *    contributors may be used to endorse or promote products derived
     24  1.4.4.2  yamt  *    from this software without specific prior written permission.
     25  1.4.4.2  yamt  *
     26  1.4.4.2  yamt  * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
     27  1.4.4.2  yamt  * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
     28  1.4.4.2  yamt  * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
     29  1.4.4.2  yamt  * PURPOSE ARE DISCLAIMED.  IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
     30  1.4.4.2  yamt  * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
     31  1.4.4.2  yamt  * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
     32  1.4.4.2  yamt  * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
     33  1.4.4.2  yamt  * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
     34  1.4.4.2  yamt  * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
     35  1.4.4.2  yamt  * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
     36  1.4.4.2  yamt  * POSSIBILITY OF SUCH DAMAGE.
     37  1.4.4.2  yamt  */
     38  1.4.4.2  yamt 
     39  1.4.4.2  yamt #ifndef _HPPA_MUTEX_H_
     40  1.4.4.2  yamt #define	_HPPA_MUTEX_H_
     41  1.4.4.2  yamt 
     42  1.4.4.2  yamt /*
     43  1.4.4.2  yamt  * The HPPA mutex implementation is troublesome, because HPPA lacks
     44  1.4.4.2  yamt  * a compare-and-set operation, yet there are many SMP HPPA machines
     45  1.4.4.2  yamt  * in circulation.  SMP for spin mutexes is easy - we don't need to
     46  1.4.4.2  yamt  * know who owns the lock.  For adaptive mutexes, we need an owner
     47  1.4.4.2  yamt  * field and additional interlock
     48  1.4.4.2  yamt  */
     49  1.4.4.2  yamt #ifndef __ASSEMBLER__
     50  1.4.4.2  yamt struct kmutex {
     51  1.4.4.2  yamt 	union {
     52  1.4.4.2  yamt 		/*
     53  1.4.4.2  yamt 		 * Only the low 4 bytes of the lock will be used by
     54  1.4.4.2  yamt 		 * __cpu_simple_lock(), but it must be aligned on a
     55  1.4.4.2  yamt 		 * 16-byte boundary.  See hppa/lock.h
     56  1.4.4.2  yamt 		 */
     57  1.4.4.2  yamt #ifdef __MUTEX_PRIVATE
     58  1.4.4.2  yamt 		__cpu_simple_lock_t	mtxu_lock;		/* 0-15 */
     59  1.4.4.2  yamt 		struct {
     60  1.4.4.2  yamt 			volatile uint32_t	mtxs_lockword;	/* 0-3 */
     61  1.4.4.2  yamt 			volatile uint32_t	mtxs_owner;	/* 4-7 */
     62  1.4.4.2  yamt 			ipl_cookie_t		mtxs_ipl;	/* 8-11 */
     63  1.4.4.2  yamt 			volatile uint8_t	mtxs_waiters;	/* 12 */
     64  1.4.4.2  yamt 
     65  1.4.4.2  yamt 			/* For LOCKDEBUG */
     66  1.4.4.2  yamt 			uint8_t			mtxs_id[3];	/* 13-15 */
     67  1.4.4.2  yamt 		} s;
     68  1.4.4.2  yamt #endif
     69  1.4.4.2  yamt 		uint8_t			mtxu_pad[16];		/* 0-15 */
     70  1.4.4.2  yamt 	} u;
     71  1.4.4.2  yamt } __aligned (16);
     72  1.4.4.2  yamt #endif
     73  1.4.4.2  yamt 
     74  1.4.4.2  yamt #ifdef __MUTEX_PRIVATE
     75  1.4.4.2  yamt 
     76  1.4.4.2  yamt #define	__HAVE_MUTEX_STUBS	1
     77  1.4.4.2  yamt 
     78  1.4.4.2  yamt #define	mtx_lock	u.mtxu_lock
     79  1.4.4.2  yamt #define	mtx_owner	u.s.mtxs_owner
     80  1.4.4.2  yamt #define	mtx_ipl		u.s.mtxs_ipl
     81  1.4.4.2  yamt #define	mtx_waiters	u.s.mtxs_waiters
     82  1.4.4.2  yamt #define	mtx_id		u.s.mtxs_id
     83  1.4.4.2  yamt 
     84  1.4.4.2  yamt /* Magic constants for mtx_owner */
     85  1.4.4.2  yamt #define	MUTEX_ADAPTIVE_UNOWNED		0xffffff00
     86  1.4.4.2  yamt #define	MUTEX_SPIN_FLAG			0xffffff10
     87  1.4.4.2  yamt #define	MUTEX_UNOWNED_OR_SPIN(x)	(((x) & 0xffffffef) == 0xffffff00)
     88  1.4.4.2  yamt 
     89  1.4.4.2  yamt #ifndef __ASSEMBLER__
     90  1.4.4.2  yamt 
     91  1.4.4.2  yamt static inline uintptr_t
     92  1.4.4.2  yamt MUTEX_OWNER(uintptr_t owner)
     93  1.4.4.2  yamt {
     94  1.4.4.2  yamt 	return owner;
     95  1.4.4.2  yamt }
     96  1.4.4.2  yamt 
     97  1.4.4.2  yamt static inline int
     98  1.4.4.2  yamt MUTEX_OWNED(uintptr_t owner)
     99  1.4.4.2  yamt {
    100  1.4.4.2  yamt 	return owner != MUTEX_ADAPTIVE_UNOWNED;
    101  1.4.4.2  yamt }
    102  1.4.4.2  yamt 
    103  1.4.4.2  yamt static inline int
    104  1.4.4.2  yamt MUTEX_SET_WAITERS(kmutex_t *mtx, uintptr_t owner)
    105  1.4.4.2  yamt {
    106  1.4.4.2  yamt 	mb_write();
    107  1.4.4.2  yamt 	mtx->mtx_waiters = 1;
    108  1.4.4.2  yamt 	mb_memory();
    109  1.4.4.2  yamt 	return mtx->mtx_owner != MUTEX_ADAPTIVE_UNOWNED;
    110  1.4.4.2  yamt }
    111  1.4.4.2  yamt 
    112  1.4.4.2  yamt static inline int
    113  1.4.4.2  yamt MUTEX_HAS_WAITERS(volatile kmutex_t *mtx)
    114  1.4.4.2  yamt {
    115  1.4.4.2  yamt 	return mtx->mtx_waiters != 0;
    116  1.4.4.2  yamt }
    117  1.4.4.2  yamt 
    118  1.4.4.2  yamt static inline void
    119  1.4.4.2  yamt MUTEX_INITIALIZE_SPIN(kmutex_t *mtx, u_int id, int ipl)
    120  1.4.4.2  yamt {
    121  1.4.4.2  yamt 	mtx->mtx_ipl = makeiplcookie(ipl);
    122  1.4.4.2  yamt 	mtx->mtx_id[0] = (uint8_t)id;
    123  1.4.4.2  yamt 	mtx->mtx_id[1] = (uint8_t)(id >> 8);
    124  1.4.4.2  yamt 	mtx->mtx_id[2] = (uint8_t)(id >> 16);
    125  1.4.4.2  yamt 	mtx->mtx_owner = MUTEX_SPIN_FLAG;
    126  1.4.4.2  yamt 	__cpu_simple_lock_init(&mtx->mtx_lock);
    127  1.4.4.2  yamt }
    128  1.4.4.2  yamt 
    129  1.4.4.2  yamt static inline void
    130  1.4.4.2  yamt MUTEX_INITIALIZE_ADAPTIVE(kmutex_t *mtx, u_int id)
    131  1.4.4.2  yamt {
    132  1.4.4.2  yamt 	mtx->mtx_id[0] = (uint8_t)id;
    133  1.4.4.2  yamt 	mtx->mtx_id[1] = (uint8_t)(id >> 8);
    134  1.4.4.2  yamt 	mtx->mtx_id[2] = (uint8_t)(id >> 16);
    135  1.4.4.2  yamt 	mtx->mtx_owner = MUTEX_ADAPTIVE_UNOWNED;
    136  1.4.4.2  yamt 	__cpu_simple_lock_init(&mtx->mtx_lock);
    137  1.4.4.2  yamt }
    138  1.4.4.2  yamt 
    139  1.4.4.2  yamt static inline void
    140  1.4.4.2  yamt MUTEX_DESTROY(kmutex_t *mtx)
    141  1.4.4.2  yamt {
    142  1.4.4.2  yamt 	mtx->mtx_owner = 0xffffffff;
    143  1.4.4.2  yamt 	mtx->mtx_id[0] = 0xff;
    144  1.4.4.2  yamt 	mtx->mtx_id[1] = 0xff;
    145  1.4.4.2  yamt 	mtx->mtx_id[2] = 0xff;
    146  1.4.4.2  yamt }
    147  1.4.4.2  yamt 
    148  1.4.4.2  yamt static inline u_int
    149  1.4.4.2  yamt MUTEX_GETID(kmutex_t *mtx)
    150  1.4.4.2  yamt {
    151  1.4.4.2  yamt 	return (u_int)mtx->mtx_id[0] |
    152  1.4.4.2  yamt 	    ((u_int)mtx->mtx_id[1] << 8) |
    153  1.4.4.2  yamt 	    ((u_int)mtx->mtx_id[2] << 16);
    154  1.4.4.2  yamt }
    155  1.4.4.2  yamt 
    156  1.4.4.2  yamt static inline int
    157  1.4.4.2  yamt MUTEX_SPIN_P(volatile kmutex_t *mtx)
    158  1.4.4.2  yamt {
    159  1.4.4.2  yamt 	return mtx->mtx_owner == MUTEX_SPIN_FLAG;
    160  1.4.4.2  yamt }
    161  1.4.4.2  yamt 
    162  1.4.4.2  yamt static inline int
    163  1.4.4.2  yamt MUTEX_ADAPTIVE_P(volatile kmutex_t *mtx)
    164  1.4.4.2  yamt {
    165  1.4.4.2  yamt 	return mtx->mtx_owner != MUTEX_SPIN_FLAG;
    166  1.4.4.2  yamt }
    167  1.4.4.2  yamt 
    168  1.4.4.2  yamt /* Acquire an adaptive mutex */
    169  1.4.4.2  yamt static inline int
    170  1.4.4.2  yamt MUTEX_ACQUIRE(kmutex_t *mtx, uintptr_t curthread)
    171  1.4.4.2  yamt {
    172  1.4.4.2  yamt 	if (!__cpu_simple_lock_try(&mtx->mtx_lock))
    173  1.4.4.2  yamt 		return 0;
    174  1.4.4.2  yamt 	mtx->mtx_owner = curthread;
    175  1.4.4.2  yamt 	return 1;
    176  1.4.4.2  yamt }
    177  1.4.4.2  yamt 
    178  1.4.4.2  yamt /* Release an adaptive mutex */
    179  1.4.4.2  yamt static inline void
    180  1.4.4.2  yamt MUTEX_RELEASE(kmutex_t *mtx)
    181  1.4.4.2  yamt {
    182  1.4.4.2  yamt 	mtx->mtx_owner = MUTEX_ADAPTIVE_UNOWNED;
    183  1.4.4.2  yamt 	__cpu_simple_unlock(&mtx->mtx_lock);
    184  1.4.4.2  yamt 	mtx->mtx_waiters = 0;
    185  1.4.4.2  yamt }
    186  1.4.4.2  yamt 
    187  1.4.4.2  yamt static inline void
    188  1.4.4.2  yamt MUTEX_CLEAR_WAITERS(kmutex_t *mtx)
    189  1.4.4.2  yamt {
    190  1.4.4.2  yamt 	mtx->mtx_waiters = 0;
    191  1.4.4.2  yamt }
    192  1.4.4.2  yamt 
    193  1.4.4.2  yamt #endif	/* __ASSEMBLER__ */
    194  1.4.4.2  yamt 
    195  1.4.4.2  yamt #endif	/* __MUTEX_PRIVATE */
    196  1.4.4.2  yamt 
    197  1.4.4.2  yamt #endif /* _HPPA_MUTEX_H_ */
    198