Home | History | Annotate | Line # | Download | only in librumpuser
      1  1.46   ozaki /*	$NetBSD: rumpuser_pth.c,v 1.46 2017/12/27 09:01:53 ozaki-r Exp $	*/
      2   1.1   pooka 
      3   1.1   pooka /*
      4   1.1   pooka  * Copyright (c) 2007-2010 Antti Kantee.  All Rights Reserved.
      5   1.1   pooka  *
      6   1.1   pooka  * Redistribution and use in source and binary forms, with or without
      7   1.1   pooka  * modification, are permitted provided that the following conditions
      8   1.1   pooka  * are met:
      9   1.1   pooka  * 1. Redistributions of source code must retain the above copyright
     10   1.1   pooka  *    notice, this list of conditions and the following disclaimer.
     11   1.1   pooka  * 2. Redistributions in binary form must reproduce the above copyright
     12   1.1   pooka  *    notice, this list of conditions and the following disclaimer in the
     13   1.1   pooka  *    documentation and/or other materials provided with the distribution.
     14   1.1   pooka  *
     15   1.1   pooka  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
     16   1.1   pooka  * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
     17   1.1   pooka  * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
     18   1.1   pooka  * DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
     19   1.1   pooka  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     20   1.1   pooka  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
     21   1.1   pooka  * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     22   1.1   pooka  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     23   1.1   pooka  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     24   1.1   pooka  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     25   1.1   pooka  * SUCH DAMAGE.
     26   1.1   pooka  */
     27   1.1   pooka 
     28   1.8   pooka #include "rumpuser_port.h"
     29   1.8   pooka 
     30   1.1   pooka #if !defined(lint)
     31  1.46   ozaki __RCSID("$NetBSD: rumpuser_pth.c,v 1.46 2017/12/27 09:01:53 ozaki-r Exp $");
     32   1.1   pooka #endif /* !lint */
     33   1.1   pooka 
     34  1.23   pooka #include <sys/queue.h>
     35  1.43   pooka 
     36  1.43   pooka #if defined(HAVE_SYS_ATOMIC_H)
     37  1.32   rmind #include <sys/atomic.h>
     38  1.32   rmind #endif
     39  1.23   pooka 
     40   1.1   pooka #include <assert.h>
     41   1.1   pooka #include <errno.h>
     42   1.8   pooka #include <fcntl.h>
     43   1.1   pooka #include <pthread.h>
     44   1.1   pooka #include <stdlib.h>
     45   1.1   pooka #include <stdio.h>
     46   1.1   pooka #include <string.h>
     47   1.1   pooka #include <stdint.h>
     48   1.1   pooka #include <unistd.h>
     49   1.1   pooka 
     50   1.1   pooka #include <rump/rumpuser.h>
     51   1.1   pooka 
     52   1.1   pooka #include "rumpuser_int.h"
     53   1.1   pooka 
     54   1.1   pooka int
     55   1.3   pooka rumpuser_thread_create(void *(*f)(void *), void *arg, const char *thrname,
     56  1.21   pooka 	int joinable, int priority, int cpuidx, void **ptcookie)
     57   1.1   pooka {
     58   1.1   pooka 	pthread_t ptid;
     59   1.3   pooka 	pthread_t *ptidp;
     60   1.3   pooka 	pthread_attr_t pattr;
     61  1.31   pooka 	int rv, i;
     62   1.1   pooka 
     63   1.3   pooka 	if ((rv = pthread_attr_init(&pattr)) != 0)
     64   1.3   pooka 		return rv;
     65   1.3   pooka 
     66   1.3   pooka 	if (joinable) {
     67   1.3   pooka 		NOFAIL(ptidp = malloc(sizeof(*ptidp)));
     68   1.3   pooka 		pthread_attr_setdetachstate(&pattr, PTHREAD_CREATE_JOINABLE);
     69   1.3   pooka 	} else {
     70   1.3   pooka 		ptidp = &ptid;
     71   1.3   pooka 		pthread_attr_setdetachstate(&pattr, PTHREAD_CREATE_DETACHED);
     72   1.3   pooka 	}
     73   1.3   pooka 
     74  1.31   pooka 	for (i = 0; i < 10; i++) {
     75  1.31   pooka 		const struct timespec ts = {0, 10*1000*1000};
     76  1.31   pooka 
     77  1.42   pooka 		rv = pthread_create(ptidp, &pattr, f, arg);
     78  1.31   pooka 		if (rv != EAGAIN)
     79  1.31   pooka 			break;
     80  1.31   pooka 		nanosleep(&ts, NULL);
     81  1.31   pooka 	}
     82  1.31   pooka 
     83  1.45   pooka #if defined(HAVE_PTHREAD_SETNAME3)
     84  1.38   pooka 	if (rv == 0 && thrname) {
     85  1.37   pooka 		pthread_setname_np(*ptidp, thrname, NULL);
     86  1.39   pooka 	}
     87  1.45   pooka #elif defined(HAVE_PTHREAD_SETNAME2)
     88  1.39   pooka 	if (rv == 0 && thrname) {
     89  1.37   pooka 		pthread_setname_np(*ptidp, thrname);
     90  1.39   pooka 	}
     91   1.1   pooka #endif
     92   1.1   pooka 
     93   1.3   pooka 	if (joinable) {
     94   1.3   pooka 		assert(ptcookie);
     95   1.3   pooka 		*ptcookie = ptidp;
     96   1.3   pooka 	}
     97   1.3   pooka 
     98   1.3   pooka 	pthread_attr_destroy(&pattr);
     99   1.3   pooka 
    100  1.20   pooka 	ET(rv);
    101   1.1   pooka }
    102   1.1   pooka 
    103   1.1   pooka __dead void
    104   1.1   pooka rumpuser_thread_exit(void)
    105   1.1   pooka {
    106   1.1   pooka 
    107  1.44   pooka 	/*
    108  1.44   pooka 	 * FIXXXME: with glibc on ARM pthread_exit() aborts because
    109  1.44   pooka 	 * it fails to unwind the stack.  In the typical case, only
    110  1.44   pooka 	 * the mountroothook thread will exit and even that's
    111  1.44   pooka 	 * conditional on vfs being present.
    112  1.44   pooka 	 */
    113  1.44   pooka #if (defined(__ARMEL__) || defined(__ARMEB__)) && defined(__GLIBC__)
    114  1.44   pooka 	for (;;)
    115  1.44   pooka 		pause();
    116  1.44   pooka #endif
    117  1.44   pooka 
    118   1.1   pooka 	pthread_exit(NULL);
    119   1.1   pooka }
    120   1.1   pooka 
    121   1.3   pooka int
    122   1.3   pooka rumpuser_thread_join(void *ptcookie)
    123   1.3   pooka {
    124   1.3   pooka 	pthread_t *pt = ptcookie;
    125   1.3   pooka 	int rv;
    126   1.3   pooka 
    127   1.3   pooka 	KLOCK_WRAP((rv = pthread_join(*pt, NULL)));
    128   1.3   pooka 	if (rv == 0)
    129   1.3   pooka 		free(pt);
    130   1.3   pooka 
    131  1.20   pooka 	ET(rv);
    132   1.3   pooka }
    133   1.3   pooka 
    134  1.26   pooka struct rumpuser_mtx {
    135  1.26   pooka 	pthread_mutex_t pthmtx;
    136  1.26   pooka 	struct lwp *owner;
    137  1.26   pooka 	int flags;
    138  1.26   pooka };
    139  1.26   pooka 
    140   1.1   pooka void
    141  1.43   pooka rumpuser_mutex_init(struct rumpuser_mtx **mtxp, int flags)
    142   1.1   pooka {
    143  1.43   pooka 	struct rumpuser_mtx *mtx;
    144   1.1   pooka 	pthread_mutexattr_t att;
    145  1.43   pooka 	size_t allocsz;
    146   1.1   pooka 
    147  1.43   pooka 	allocsz = (sizeof(*mtx)+RUMPUSER_LOCKALIGN) & ~(RUMPUSER_LOCKALIGN-1);
    148  1.43   pooka 	NOFAIL(mtx = aligned_alloc(RUMPUSER_LOCKALIGN, allocsz));
    149   1.1   pooka 
    150   1.1   pooka 	pthread_mutexattr_init(&att);
    151   1.1   pooka 	pthread_mutexattr_settype(&att, PTHREAD_MUTEX_ERRORCHECK);
    152  1.43   pooka 	NOFAIL_ERRNO(pthread_mutex_init(&mtx->pthmtx, &att));
    153   1.1   pooka 	pthread_mutexattr_destroy(&att);
    154   1.1   pooka 
    155  1.43   pooka 	mtx->owner = NULL;
    156  1.15   pooka 	assert(flags != 0);
    157  1.43   pooka 	mtx->flags = flags;
    158  1.43   pooka 
    159  1.43   pooka 	*mtxp = mtx;
    160   1.4   pooka }
    161   1.4   pooka 
    162  1.46   ozaki int
    163  1.46   ozaki rumpuser_mutex_spin_p(struct rumpuser_mtx *mtx)
    164  1.46   ozaki {
    165  1.46   ozaki 
    166  1.46   ozaki 	return (mtx->flags & RUMPUSER_MTX_SPIN) != 0;
    167  1.46   ozaki }
    168  1.46   ozaki 
    169   1.1   pooka static void
    170   1.1   pooka mtxenter(struct rumpuser_mtx *mtx)
    171   1.1   pooka {
    172   1.1   pooka 
    173  1.15   pooka 	if (!(mtx->flags & RUMPUSER_MTX_KMUTEX))
    174   1.4   pooka 		return;
    175   1.4   pooka 
    176   1.4   pooka 	assert(mtx->owner == NULL);
    177  1.23   pooka 	mtx->owner = rumpuser_curlwp();
    178   1.1   pooka }
    179   1.1   pooka 
    180   1.1   pooka static void
    181   1.1   pooka mtxexit(struct rumpuser_mtx *mtx)
    182   1.1   pooka {
    183   1.1   pooka 
    184  1.15   pooka 	if (!(mtx->flags & RUMPUSER_MTX_KMUTEX))
    185   1.4   pooka 		return;
    186   1.4   pooka 
    187   1.1   pooka 	assert(mtx->owner != NULL);
    188   1.4   pooka 	mtx->owner = NULL;
    189   1.1   pooka }
    190   1.1   pooka 
    191   1.1   pooka void
    192   1.1   pooka rumpuser_mutex_enter(struct rumpuser_mtx *mtx)
    193   1.1   pooka {
    194   1.1   pooka 
    195  1.15   pooka 	if (mtx->flags & RUMPUSER_MTX_SPIN) {
    196  1.13   pooka 		rumpuser_mutex_enter_nowrap(mtx);
    197  1.13   pooka 		return;
    198  1.13   pooka 	}
    199  1.13   pooka 
    200  1.15   pooka 	assert(mtx->flags & RUMPUSER_MTX_KMUTEX);
    201   1.1   pooka 	if (pthread_mutex_trylock(&mtx->pthmtx) != 0)
    202   1.1   pooka 		KLOCK_WRAP(NOFAIL_ERRNO(pthread_mutex_lock(&mtx->pthmtx)));
    203   1.1   pooka 	mtxenter(mtx);
    204   1.1   pooka }
    205   1.1   pooka 
    206   1.1   pooka void
    207   1.1   pooka rumpuser_mutex_enter_nowrap(struct rumpuser_mtx *mtx)
    208   1.1   pooka {
    209   1.1   pooka 
    210  1.15   pooka 	assert(mtx->flags & RUMPUSER_MTX_SPIN);
    211   1.1   pooka 	NOFAIL_ERRNO(pthread_mutex_lock(&mtx->pthmtx));
    212   1.1   pooka 	mtxenter(mtx);
    213   1.1   pooka }
    214   1.1   pooka 
    215   1.1   pooka int
    216   1.1   pooka rumpuser_mutex_tryenter(struct rumpuser_mtx *mtx)
    217   1.1   pooka {
    218   1.1   pooka 	int rv;
    219   1.1   pooka 
    220   1.1   pooka 	rv = pthread_mutex_trylock(&mtx->pthmtx);
    221   1.1   pooka 	if (rv == 0) {
    222   1.1   pooka 		mtxenter(mtx);
    223   1.1   pooka 	}
    224   1.1   pooka 
    225  1.20   pooka 	ET(rv);
    226   1.1   pooka }
    227   1.1   pooka 
    228   1.1   pooka void
    229   1.1   pooka rumpuser_mutex_exit(struct rumpuser_mtx *mtx)
    230   1.1   pooka {
    231   1.1   pooka 
    232   1.1   pooka 	mtxexit(mtx);
    233   1.1   pooka 	NOFAIL_ERRNO(pthread_mutex_unlock(&mtx->pthmtx));
    234   1.1   pooka }
    235   1.1   pooka 
    236   1.1   pooka void
    237   1.1   pooka rumpuser_mutex_destroy(struct rumpuser_mtx *mtx)
    238   1.1   pooka {
    239   1.1   pooka 
    240   1.1   pooka 	NOFAIL_ERRNO(pthread_mutex_destroy(&mtx->pthmtx));
    241   1.1   pooka 	free(mtx);
    242   1.1   pooka }
    243   1.1   pooka 
    244  1.19   pooka void
    245  1.19   pooka rumpuser_mutex_owner(struct rumpuser_mtx *mtx, struct lwp **lp)
    246   1.1   pooka {
    247   1.1   pooka 
    248  1.15   pooka 	if (__predict_false(!(mtx->flags & RUMPUSER_MTX_KMUTEX))) {
    249   1.4   pooka 		printf("panic: rumpuser_mutex_held unsupported on non-kmtx\n");
    250   1.4   pooka 		abort();
    251   1.4   pooka 	}
    252   1.4   pooka 
    253  1.19   pooka 	*lp = mtx->owner;
    254   1.1   pooka }
    255   1.1   pooka 
    256  1.26   pooka /*
    257  1.27   pooka  * rwlocks.  these are mostly simple, except that NetBSD wants to
    258  1.27   pooka  * support something called downgrade, which means we need to swap
    259  1.27   pooka  * our exclusive lock for a shared lock.  to accommodate this,
    260  1.27   pooka  * we need to check *after* acquiring a lock in case someone was
    261  1.27   pooka  * downgrading it.  if so, we couldn't actually have it and maybe
    262  1.27   pooka  * need to retry later.
    263  1.26   pooka  */
    264  1.26   pooka 
    265  1.26   pooka struct rumpuser_rw {
    266  1.26   pooka 	pthread_rwlock_t pthrw;
    267  1.40  justin #if !defined(__APPLE__) && !defined(__ANDROID__)
    268  1.33   rmind 	char pad[64 - sizeof(pthread_rwlock_t)];
    269  1.26   pooka 	pthread_spinlock_t spin;
    270  1.34   rmind #endif
    271  1.32   rmind 	unsigned int readers;
    272  1.26   pooka 	struct lwp *writer;
    273  1.27   pooka 	int downgrade; /* someone is downgrading (hopefully lock holder ;) */
    274  1.26   pooka };
    275  1.26   pooka 
    276  1.27   pooka static int
    277  1.27   pooka rw_amwriter(struct rumpuser_rw *rw)
    278  1.27   pooka {
    279  1.27   pooka 
    280  1.32   rmind 	return rw->writer == rumpuser_curlwp() && rw->readers == (unsigned)-1;
    281  1.27   pooka }
    282  1.27   pooka 
    283  1.27   pooka static int
    284  1.27   pooka rw_nreaders(struct rumpuser_rw *rw)
    285  1.27   pooka {
    286  1.32   rmind 	unsigned nreaders = rw->readers;
    287  1.27   pooka 
    288  1.32   rmind 	return nreaders != (unsigned)-1 ? nreaders : 0;
    289  1.27   pooka }
    290  1.27   pooka 
    291  1.27   pooka static int
    292  1.27   pooka rw_setwriter(struct rumpuser_rw *rw, int retry)
    293  1.27   pooka {
    294  1.27   pooka 
    295  1.27   pooka 	/*
    296  1.27   pooka 	 * Don't need the spinlock here, we already have an
    297  1.27   pooka 	 * exclusive lock and "downgrade" is stable until complete.
    298  1.27   pooka 	 */
    299  1.27   pooka 	if (rw->downgrade) {
    300  1.27   pooka 		pthread_rwlock_unlock(&rw->pthrw);
    301  1.27   pooka 		if (retry) {
    302  1.27   pooka 			struct timespec ts;
    303  1.27   pooka 
    304  1.27   pooka 			/* portable yield, essentially */
    305  1.27   pooka 			ts.tv_sec = 0;
    306  1.27   pooka 			ts.tv_nsec = 1;
    307  1.27   pooka 			KLOCK_WRAP(nanosleep(&ts, NULL));
    308  1.27   pooka 		}
    309  1.27   pooka 		return EBUSY;
    310  1.27   pooka 	}
    311  1.27   pooka 	assert(rw->readers == 0);
    312  1.27   pooka 	rw->writer = rumpuser_curlwp();
    313  1.32   rmind 	rw->readers = (unsigned)-1;
    314  1.27   pooka 	return 0;
    315  1.27   pooka }
    316  1.27   pooka 
    317  1.27   pooka static void
    318  1.27   pooka rw_clearwriter(struct rumpuser_rw *rw)
    319  1.27   pooka {
    320  1.27   pooka 
    321  1.27   pooka 	assert(rw_amwriter(rw));
    322  1.27   pooka 	rw->readers = 0;
    323  1.27   pooka 	rw->writer = NULL;
    324  1.27   pooka }
    325  1.27   pooka 
    326  1.33   rmind static inline void
    327  1.27   pooka rw_readup(struct rumpuser_rw *rw)
    328  1.27   pooka {
    329  1.27   pooka 
    330  1.40  justin #if defined(__NetBSD__) || defined(__APPLE__) || defined(__ANDROID__)
    331  1.32   rmind 	atomic_inc_uint(&rw->readers);
    332  1.32   rmind #else
    333  1.27   pooka 	pthread_spin_lock(&rw->spin);
    334  1.27   pooka 	++rw->readers;
    335  1.27   pooka 	pthread_spin_unlock(&rw->spin);
    336  1.32   rmind #endif
    337  1.27   pooka }
    338  1.27   pooka 
    339  1.33   rmind static inline void
    340  1.27   pooka rw_readdown(struct rumpuser_rw *rw)
    341  1.27   pooka {
    342  1.27   pooka 
    343  1.40  justin #if defined(__NetBSD__) || defined(__APPLE__) || defined(__ANDROID__)
    344  1.32   rmind 	atomic_dec_uint(&rw->readers);
    345  1.32   rmind #else
    346  1.27   pooka 	pthread_spin_lock(&rw->spin);
    347  1.27   pooka 	assert(rw->readers > 0);
    348  1.27   pooka 	--rw->readers;
    349  1.27   pooka 	pthread_spin_unlock(&rw->spin);
    350  1.32   rmind #endif
    351  1.27   pooka }
    352  1.26   pooka 
    353   1.1   pooka void
    354  1.43   pooka rumpuser_rw_init(struct rumpuser_rw **rwp)
    355   1.1   pooka {
    356  1.43   pooka 	struct rumpuser_rw *rw;
    357  1.43   pooka 	size_t allocsz;
    358   1.1   pooka 
    359  1.43   pooka 	allocsz = (sizeof(*rw)+RUMPUSER_LOCKALIGN) & ~(RUMPUSER_LOCKALIGN-1);
    360  1.43   pooka 
    361  1.43   pooka 	NOFAIL(rw = aligned_alloc(RUMPUSER_LOCKALIGN, allocsz));
    362  1.43   pooka 	NOFAIL_ERRNO(pthread_rwlock_init(&rw->pthrw, NULL));
    363  1.40  justin #if !defined(__APPLE__) && !defined(__ANDROID__)
    364  1.43   pooka 	NOFAIL_ERRNO(pthread_spin_init(&rw->spin, PTHREAD_PROCESS_PRIVATE));
    365  1.34   rmind #endif
    366  1.43   pooka 	rw->readers = 0;
    367  1.43   pooka 	rw->writer = NULL;
    368  1.43   pooka 	rw->downgrade = 0;
    369  1.43   pooka 
    370  1.43   pooka 	*rwp = rw;
    371   1.1   pooka }
    372   1.1   pooka 
    373   1.1   pooka void
    374  1.30   pooka rumpuser_rw_enter(int enum_rumprwlock, struct rumpuser_rw *rw)
    375   1.1   pooka {
    376  1.30   pooka 	enum rumprwlock lk = enum_rumprwlock;
    377   1.1   pooka 
    378  1.25   pooka 	switch (lk) {
    379  1.25   pooka 	case RUMPUSER_RW_WRITER:
    380  1.27   pooka 		do {
    381  1.27   pooka 			if (pthread_rwlock_trywrlock(&rw->pthrw) != 0)
    382  1.27   pooka 				KLOCK_WRAP(NOFAIL_ERRNO(
    383  1.27   pooka 				    pthread_rwlock_wrlock(&rw->pthrw)));
    384  1.27   pooka 		} while (rw_setwriter(rw, 1) != 0);
    385  1.25   pooka 		break;
    386  1.25   pooka 	case RUMPUSER_RW_READER:
    387   1.1   pooka 		if (pthread_rwlock_tryrdlock(&rw->pthrw) != 0)
    388   1.1   pooka 			KLOCK_WRAP(NOFAIL_ERRNO(
    389   1.1   pooka 			    pthread_rwlock_rdlock(&rw->pthrw)));
    390  1.27   pooka 		rw_readup(rw);
    391  1.25   pooka 		break;
    392   1.1   pooka 	}
    393   1.1   pooka }
    394   1.1   pooka 
    395   1.1   pooka int
    396  1.30   pooka rumpuser_rw_tryenter(int enum_rumprwlock, struct rumpuser_rw *rw)
    397   1.1   pooka {
    398  1.30   pooka 	enum rumprwlock lk = enum_rumprwlock;
    399   1.1   pooka 	int rv;
    400   1.1   pooka 
    401  1.25   pooka 	switch (lk) {
    402  1.25   pooka 	case RUMPUSER_RW_WRITER:
    403   1.1   pooka 		rv = pthread_rwlock_trywrlock(&rw->pthrw);
    404   1.1   pooka 		if (rv == 0)
    405  1.27   pooka 			rv = rw_setwriter(rw, 0);
    406  1.25   pooka 		break;
    407  1.25   pooka 	case RUMPUSER_RW_READER:
    408   1.1   pooka 		rv = pthread_rwlock_tryrdlock(&rw->pthrw);
    409   1.1   pooka 		if (rv == 0)
    410  1.27   pooka 			rw_readup(rw);
    411  1.25   pooka 		break;
    412  1.25   pooka 	default:
    413  1.25   pooka 		rv = EINVAL;
    414  1.25   pooka 		break;
    415   1.1   pooka 	}
    416   1.1   pooka 
    417  1.20   pooka 	ET(rv);
    418   1.1   pooka }
    419   1.1   pooka 
    420  1.25   pooka int
    421  1.25   pooka rumpuser_rw_tryupgrade(struct rumpuser_rw *rw)
    422  1.25   pooka {
    423  1.25   pooka 
    424  1.27   pooka 	/*
    425  1.27   pooka 	 * Not supported by pthreads.  Since the caller needs to
    426  1.27   pooka 	 * back off anyway to avoid deadlock, always failing
    427  1.27   pooka 	 * is correct.
    428  1.27   pooka 	 */
    429  1.25   pooka 	ET(EBUSY);
    430  1.25   pooka }
    431  1.25   pooka 
    432  1.27   pooka /*
    433  1.27   pooka  * convert from exclusive to shared lock without allowing anyone to
    434  1.27   pooka  * obtain an exclusive lock in between.  actually, might allow
    435  1.27   pooka  * someone to obtain the lock, we just don't allow that thread to
    436  1.27   pooka  * return from the hypercall with it.
    437  1.27   pooka  */
    438  1.25   pooka void
    439  1.25   pooka rumpuser_rw_downgrade(struct rumpuser_rw *rw)
    440  1.25   pooka {
    441  1.25   pooka 
    442  1.27   pooka 	assert(rw->downgrade == 0);
    443  1.27   pooka 	rw->downgrade = 1;
    444  1.27   pooka 	rumpuser_rw_exit(rw);
    445  1.25   pooka 	/*
    446  1.27   pooka 	 * though the competition can't get out of the hypervisor, it
    447  1.27   pooka 	 * might have rescheduled itself after we released the lock.
    448  1.27   pooka 	 * so need a wrap here.
    449  1.25   pooka 	 */
    450  1.27   pooka 	KLOCK_WRAP(NOFAIL_ERRNO(pthread_rwlock_rdlock(&rw->pthrw)));
    451  1.27   pooka 	rw->downgrade = 0;
    452  1.27   pooka 	rw_readup(rw);
    453  1.25   pooka }
    454  1.25   pooka 
    455   1.1   pooka void
    456   1.1   pooka rumpuser_rw_exit(struct rumpuser_rw *rw)
    457   1.1   pooka {
    458   1.1   pooka 
    459  1.27   pooka 	if (rw_nreaders(rw))
    460  1.27   pooka 		rw_readdown(rw);
    461   1.1   pooka 	else
    462  1.27   pooka 		rw_clearwriter(rw);
    463   1.1   pooka 	NOFAIL_ERRNO(pthread_rwlock_unlock(&rw->pthrw));
    464   1.1   pooka }
    465   1.1   pooka 
    466   1.1   pooka void
    467   1.1   pooka rumpuser_rw_destroy(struct rumpuser_rw *rw)
    468   1.1   pooka {
    469   1.1   pooka 
    470   1.1   pooka 	NOFAIL_ERRNO(pthread_rwlock_destroy(&rw->pthrw));
    471  1.40  justin #if !defined(__APPLE__) && ! defined(__ANDROID__)
    472   1.1   pooka 	NOFAIL_ERRNO(pthread_spin_destroy(&rw->spin));
    473  1.34   rmind #endif
    474   1.1   pooka 	free(rw);
    475   1.1   pooka }
    476   1.1   pooka 
    477  1.19   pooka void
    478  1.30   pooka rumpuser_rw_held(int enum_rumprwlock, struct rumpuser_rw *rw, int *rv)
    479   1.1   pooka {
    480  1.30   pooka 	enum rumprwlock lk = enum_rumprwlock;
    481   1.1   pooka 
    482  1.25   pooka 	switch (lk) {
    483  1.25   pooka 	case RUMPUSER_RW_WRITER:
    484  1.27   pooka 		*rv = rw_amwriter(rw);
    485  1.25   pooka 		break;
    486  1.25   pooka 	case RUMPUSER_RW_READER:
    487  1.27   pooka 		*rv = rw_nreaders(rw);
    488  1.25   pooka 		break;
    489  1.25   pooka 	}
    490   1.1   pooka }
    491   1.1   pooka 
    492  1.26   pooka /*
    493  1.26   pooka  * condvar
    494  1.26   pooka  */
    495  1.26   pooka 
    496  1.26   pooka struct rumpuser_cv {
    497  1.26   pooka 	pthread_cond_t pthcv;
    498  1.26   pooka 	int nwaiters;
    499  1.26   pooka };
    500  1.26   pooka 
    501   1.1   pooka void
    502   1.1   pooka rumpuser_cv_init(struct rumpuser_cv **cv)
    503   1.1   pooka {
    504   1.1   pooka 
    505   1.1   pooka 	NOFAIL(*cv = malloc(sizeof(struct rumpuser_cv)));
    506   1.1   pooka 	NOFAIL_ERRNO(pthread_cond_init(&((*cv)->pthcv), NULL));
    507   1.1   pooka 	(*cv)->nwaiters = 0;
    508   1.1   pooka }
    509   1.1   pooka 
    510   1.1   pooka void
    511   1.1   pooka rumpuser_cv_destroy(struct rumpuser_cv *cv)
    512   1.1   pooka {
    513   1.1   pooka 
    514   1.1   pooka 	NOFAIL_ERRNO(pthread_cond_destroy(&cv->pthcv));
    515   1.1   pooka 	free(cv);
    516   1.1   pooka }
    517   1.1   pooka 
    518  1.24   pooka static void
    519  1.24   pooka cv_unschedule(struct rumpuser_mtx *mtx, int *nlocks)
    520  1.24   pooka {
    521  1.24   pooka 
    522  1.24   pooka 	rumpkern_unsched(nlocks, mtx);
    523  1.24   pooka 	mtxexit(mtx);
    524  1.24   pooka }
    525  1.24   pooka 
    526  1.24   pooka static void
    527  1.24   pooka cv_reschedule(struct rumpuser_mtx *mtx, int nlocks)
    528  1.24   pooka {
    529  1.24   pooka 
    530  1.24   pooka 	/*
    531  1.24   pooka 	 * If the cv interlock is a spin mutex, we must first release
    532  1.24   pooka 	 * the mutex that was reacquired by pthread_cond_wait(),
    533  1.24   pooka 	 * acquire the CPU context and only then relock the mutex.
    534  1.24   pooka 	 * This is to preserve resource allocation order so that
    535  1.24   pooka 	 * we don't deadlock.  Non-spinning mutexes don't have this
    536  1.24   pooka 	 * problem since they don't use a hold-and-wait approach
    537  1.24   pooka 	 * to acquiring the mutex wrt the rump kernel CPU context.
    538  1.24   pooka 	 *
    539  1.24   pooka 	 * The more optimal solution would be to rework rumpkern_sched()
    540  1.24   pooka 	 * so that it's possible to tell the scheduler
    541  1.24   pooka 	 * "if you need to block, drop this lock first", but I'm not
    542  1.24   pooka 	 * going poking there without some numbers on how often this
    543  1.24   pooka 	 * path is taken for spin mutexes.
    544  1.24   pooka 	 */
    545  1.24   pooka 	if ((mtx->flags & (RUMPUSER_MTX_SPIN | RUMPUSER_MTX_KMUTEX)) ==
    546  1.24   pooka 	    (RUMPUSER_MTX_SPIN | RUMPUSER_MTX_KMUTEX)) {
    547  1.24   pooka 		NOFAIL_ERRNO(pthread_mutex_unlock(&mtx->pthmtx));
    548  1.24   pooka 		rumpkern_sched(nlocks, mtx);
    549  1.24   pooka 		rumpuser_mutex_enter_nowrap(mtx);
    550  1.24   pooka 	} else {
    551  1.24   pooka 		mtxenter(mtx);
    552  1.24   pooka 		rumpkern_sched(nlocks, mtx);
    553  1.24   pooka 	}
    554  1.24   pooka }
    555  1.24   pooka 
    556   1.1   pooka void
    557   1.1   pooka rumpuser_cv_wait(struct rumpuser_cv *cv, struct rumpuser_mtx *mtx)
    558   1.1   pooka {
    559   1.2   pooka 	int nlocks;
    560   1.1   pooka 
    561   1.1   pooka 	cv->nwaiters++;
    562  1.24   pooka 	cv_unschedule(mtx, &nlocks);
    563   1.2   pooka 	NOFAIL_ERRNO(pthread_cond_wait(&cv->pthcv, &mtx->pthmtx));
    564  1.24   pooka 	cv_reschedule(mtx, nlocks);
    565   1.1   pooka 	cv->nwaiters--;
    566   1.1   pooka }
    567   1.1   pooka 
    568   1.1   pooka void
    569   1.1   pooka rumpuser_cv_wait_nowrap(struct rumpuser_cv *cv, struct rumpuser_mtx *mtx)
    570   1.1   pooka {
    571   1.1   pooka 
    572   1.1   pooka 	cv->nwaiters++;
    573   1.1   pooka 	mtxexit(mtx);
    574   1.1   pooka 	NOFAIL_ERRNO(pthread_cond_wait(&cv->pthcv, &mtx->pthmtx));
    575   1.1   pooka 	mtxenter(mtx);
    576   1.1   pooka 	cv->nwaiters--;
    577   1.1   pooka }
    578   1.1   pooka 
    579   1.1   pooka int
    580   1.1   pooka rumpuser_cv_timedwait(struct rumpuser_cv *cv, struct rumpuser_mtx *mtx,
    581   1.1   pooka 	int64_t sec, int64_t nsec)
    582   1.1   pooka {
    583   1.1   pooka 	struct timespec ts;
    584   1.2   pooka 	int rv, nlocks;
    585   1.1   pooka 
    586  1.16   pooka 	/*
    587  1.16   pooka 	 * Get clock already here, just in case we will be put to sleep
    588  1.16   pooka 	 * after releasing the kernel context.
    589  1.16   pooka 	 *
    590  1.16   pooka 	 * The condition variables should use CLOCK_MONOTONIC, but since
    591  1.16   pooka 	 * that's not available everywhere, leave it for another day.
    592  1.16   pooka 	 */
    593  1.16   pooka 	clock_gettime(CLOCK_REALTIME, &ts);
    594   1.1   pooka 
    595   1.1   pooka 	cv->nwaiters++;
    596  1.24   pooka 	cv_unschedule(mtx, &nlocks);
    597  1.16   pooka 
    598  1.16   pooka 	ts.tv_sec += sec;
    599  1.16   pooka 	ts.tv_nsec += nsec;
    600  1.16   pooka 	if (ts.tv_nsec >= 1000*1000*1000) {
    601  1.16   pooka 		ts.tv_sec++;
    602  1.16   pooka 		ts.tv_nsec -= 1000*1000*1000;
    603  1.16   pooka 	}
    604   1.2   pooka 	rv = pthread_cond_timedwait(&cv->pthcv, &mtx->pthmtx, &ts);
    605  1.24   pooka 
    606  1.24   pooka 	cv_reschedule(mtx, nlocks);
    607   1.1   pooka 	cv->nwaiters--;
    608   1.1   pooka 
    609  1.20   pooka 	ET(rv);
    610   1.1   pooka }
    611   1.1   pooka 
    612   1.1   pooka void
    613   1.1   pooka rumpuser_cv_signal(struct rumpuser_cv *cv)
    614   1.1   pooka {
    615   1.1   pooka 
    616   1.1   pooka 	NOFAIL_ERRNO(pthread_cond_signal(&cv->pthcv));
    617   1.1   pooka }
    618   1.1   pooka 
    619   1.1   pooka void
    620   1.1   pooka rumpuser_cv_broadcast(struct rumpuser_cv *cv)
    621   1.1   pooka {
    622   1.1   pooka 
    623   1.1   pooka 	NOFAIL_ERRNO(pthread_cond_broadcast(&cv->pthcv));
    624   1.1   pooka }
    625   1.1   pooka 
    626  1.19   pooka void
    627  1.19   pooka rumpuser_cv_has_waiters(struct rumpuser_cv *cv, int *nwaiters)
    628   1.1   pooka {
    629   1.1   pooka 
    630  1.19   pooka 	*nwaiters = cv->nwaiters;
    631   1.1   pooka }
    632   1.1   pooka 
    633   1.1   pooka /*
    634   1.1   pooka  * curlwp
    635   1.1   pooka  */
    636   1.1   pooka 
    637  1.36  justin static pthread_key_t curlwpkey;
    638  1.26   pooka 
    639  1.23   pooka /*
    640  1.23   pooka  * the if0'd curlwp implementation is not used by this hypervisor,
    641  1.23   pooka  * but serves as test code to check that the intended usage works.
    642  1.23   pooka  */
    643  1.23   pooka #if 0
    644  1.23   pooka struct rumpuser_lwp {
    645  1.23   pooka 	struct lwp *l;
    646  1.23   pooka 	LIST_ENTRY(rumpuser_lwp) l_entries;
    647  1.23   pooka };
    648  1.23   pooka static LIST_HEAD(, rumpuser_lwp) lwps = LIST_HEAD_INITIALIZER(lwps);
    649  1.23   pooka static pthread_mutex_t lwplock = PTHREAD_MUTEX_INITIALIZER;
    650  1.23   pooka 
    651   1.1   pooka void
    652  1.23   pooka rumpuser_curlwpop(enum rumplwpop op, struct lwp *l)
    653   1.1   pooka {
    654  1.23   pooka 	struct rumpuser_lwp *rl, *rliter;
    655   1.1   pooka 
    656  1.23   pooka 	switch (op) {
    657  1.23   pooka 	case RUMPUSER_LWP_CREATE:
    658  1.23   pooka 		rl = malloc(sizeof(*rl));
    659  1.23   pooka 		rl->l = l;
    660  1.23   pooka 		pthread_mutex_lock(&lwplock);
    661  1.23   pooka 		LIST_FOREACH(rliter, &lwps, l_entries) {
    662  1.23   pooka 			if (rliter->l == l) {
    663  1.23   pooka 				fprintf(stderr, "LWP_CREATE: %p exists\n", l);
    664  1.23   pooka 				abort();
    665  1.23   pooka 			}
    666  1.23   pooka 		}
    667  1.23   pooka 		LIST_INSERT_HEAD(&lwps, rl, l_entries);
    668  1.23   pooka 		pthread_mutex_unlock(&lwplock);
    669  1.23   pooka 		break;
    670  1.23   pooka 	case RUMPUSER_LWP_DESTROY:
    671  1.23   pooka 		pthread_mutex_lock(&lwplock);
    672  1.23   pooka 		LIST_FOREACH(rl, &lwps, l_entries) {
    673  1.23   pooka 			if (rl->l == l)
    674  1.23   pooka 				break;
    675  1.23   pooka 		}
    676  1.23   pooka 		if (!rl) {
    677  1.23   pooka 			fprintf(stderr, "LWP_DESTROY: %p does not exist\n", l);
    678  1.23   pooka 			abort();
    679  1.23   pooka 		}
    680  1.23   pooka 		LIST_REMOVE(rl, l_entries);
    681  1.23   pooka 		pthread_mutex_unlock(&lwplock);
    682  1.23   pooka 		free(rl);
    683  1.23   pooka 		break;
    684  1.23   pooka 	case RUMPUSER_LWP_SET:
    685  1.29   pooka 		assert(pthread_getspecific(curlwpkey) == NULL && l != NULL);
    686  1.23   pooka 
    687  1.29   pooka 		pthread_mutex_lock(&lwplock);
    688  1.29   pooka 		LIST_FOREACH(rl, &lwps, l_entries) {
    689  1.29   pooka 			if (rl->l == l)
    690  1.29   pooka 				break;
    691  1.29   pooka 		}
    692  1.29   pooka 		if (!rl) {
    693  1.29   pooka 			fprintf(stderr,
    694  1.29   pooka 			    "LWP_SET: %p does not exist\n", l);
    695  1.29   pooka 			abort();
    696  1.23   pooka 		}
    697  1.29   pooka 		pthread_mutex_unlock(&lwplock);
    698  1.23   pooka 
    699  1.23   pooka 		pthread_setspecific(curlwpkey, rl);
    700  1.23   pooka 		break;
    701  1.29   pooka 	case RUMPUSER_LWP_CLEAR:
    702  1.29   pooka 		assert(((struct rumpuser_lwp *)
    703  1.29   pooka 		    pthread_getspecific(curlwpkey))->l == l);
    704  1.29   pooka 		pthread_setspecific(curlwpkey, NULL);
    705  1.29   pooka 		break;
    706  1.23   pooka 	}
    707   1.1   pooka }
    708   1.1   pooka 
    709   1.1   pooka struct lwp *
    710  1.23   pooka rumpuser_curlwp(void)
    711  1.23   pooka {
    712  1.23   pooka 	struct rumpuser_lwp *rl;
    713  1.23   pooka 
    714  1.23   pooka 	rl = pthread_getspecific(curlwpkey);
    715  1.23   pooka 	return rl ? rl->l : NULL;
    716  1.23   pooka }
    717  1.23   pooka 
    718  1.23   pooka #else
    719  1.23   pooka 
    720  1.23   pooka void
    721  1.30   pooka rumpuser_curlwpop(int enum_rumplwpop, struct lwp *l)
    722  1.23   pooka {
    723  1.30   pooka 	enum rumplwpop op = enum_rumplwpop;
    724  1.23   pooka 
    725  1.23   pooka 	switch (op) {
    726  1.23   pooka 	case RUMPUSER_LWP_CREATE:
    727  1.23   pooka 		break;
    728  1.23   pooka 	case RUMPUSER_LWP_DESTROY:
    729  1.23   pooka 		break;
    730  1.23   pooka 	case RUMPUSER_LWP_SET:
    731  1.36  justin 		assert(pthread_getspecific(curlwpkey) == NULL);
    732  1.36  justin 		pthread_setspecific(curlwpkey, l);
    733  1.23   pooka 		break;
    734  1.29   pooka 	case RUMPUSER_LWP_CLEAR:
    735  1.36  justin 		assert(pthread_getspecific(curlwpkey) == l);
    736  1.36  justin 		pthread_setspecific(curlwpkey, NULL);
    737  1.29   pooka 		break;
    738  1.23   pooka 	}
    739  1.23   pooka }
    740  1.23   pooka 
    741  1.23   pooka struct lwp *
    742  1.23   pooka rumpuser_curlwp(void)
    743   1.1   pooka {
    744   1.1   pooka 
    745  1.36  justin 	return pthread_getspecific(curlwpkey);
    746   1.1   pooka }
    747  1.23   pooka #endif
    748  1.26   pooka 
    749  1.26   pooka 
    750  1.26   pooka void
    751  1.26   pooka rumpuser__thrinit(void)
    752  1.26   pooka {
    753  1.36  justin 	pthread_key_create(&curlwpkey, NULL);
    754  1.26   pooka }
    755