Home | History | Annotate | Line # | Download | only in librumpuser
rumpuser_pth.c revision 1.29
      1  1.29  pooka /*	$NetBSD: rumpuser_pth.c,v 1.29 2013/05/15 14:07:26 pooka Exp $	*/
      2   1.1  pooka 
      3   1.1  pooka /*
      4   1.1  pooka  * Copyright (c) 2007-2010 Antti Kantee.  All Rights Reserved.
      5   1.1  pooka  *
      6   1.1  pooka  * Redistribution and use in source and binary forms, with or without
      7   1.1  pooka  * modification, are permitted provided that the following conditions
      8   1.1  pooka  * are met:
      9   1.1  pooka  * 1. Redistributions of source code must retain the above copyright
     10   1.1  pooka  *    notice, this list of conditions and the following disclaimer.
     11   1.1  pooka  * 2. Redistributions in binary form must reproduce the above copyright
     12   1.1  pooka  *    notice, this list of conditions and the following disclaimer in the
     13   1.1  pooka  *    documentation and/or other materials provided with the distribution.
     14   1.1  pooka  *
     15   1.1  pooka  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS
     16   1.1  pooka  * OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
     17   1.1  pooka  * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
     18   1.1  pooka  * DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
     19   1.1  pooka  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     20   1.1  pooka  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
     21   1.1  pooka  * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     22   1.1  pooka  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     23   1.1  pooka  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     24   1.1  pooka  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     25   1.1  pooka  * SUCH DAMAGE.
     26   1.1  pooka  */
     27   1.1  pooka 
     28   1.8  pooka #include "rumpuser_port.h"
     29   1.8  pooka 
     30   1.1  pooka #if !defined(lint)
     31  1.29  pooka __RCSID("$NetBSD: rumpuser_pth.c,v 1.29 2013/05/15 14:07:26 pooka Exp $");
     32   1.1  pooka #endif /* !lint */
     33   1.1  pooka 
     34  1.23  pooka #include <sys/queue.h>
     35  1.23  pooka 
     36   1.1  pooka #include <assert.h>
     37   1.1  pooka #include <errno.h>
     38   1.8  pooka #include <fcntl.h>
     39   1.1  pooka #include <pthread.h>
     40   1.1  pooka #include <stdlib.h>
     41   1.1  pooka #include <stdio.h>
     42   1.1  pooka #include <string.h>
     43   1.1  pooka #include <stdint.h>
     44   1.1  pooka #include <unistd.h>
     45   1.1  pooka 
     46   1.1  pooka #include <rump/rumpuser.h>
     47   1.1  pooka 
     48   1.1  pooka #include "rumpuser_int.h"
     49   1.1  pooka 
     50   1.1  pooka int
     51   1.3  pooka rumpuser_thread_create(void *(*f)(void *), void *arg, const char *thrname,
     52  1.21  pooka 	int joinable, int priority, int cpuidx, void **ptcookie)
     53   1.1  pooka {
     54   1.1  pooka 	pthread_t ptid;
     55   1.3  pooka 	pthread_t *ptidp;
     56   1.3  pooka 	pthread_attr_t pattr;
     57   1.1  pooka 	int rv;
     58   1.1  pooka 
     59   1.3  pooka 	if ((rv = pthread_attr_init(&pattr)) != 0)
     60   1.3  pooka 		return rv;
     61   1.3  pooka 
     62   1.3  pooka 	if (joinable) {
     63   1.3  pooka 		NOFAIL(ptidp = malloc(sizeof(*ptidp)));
     64   1.3  pooka 		pthread_attr_setdetachstate(&pattr, PTHREAD_CREATE_JOINABLE);
     65   1.3  pooka 	} else {
     66   1.3  pooka 		ptidp = &ptid;
     67   1.3  pooka 		pthread_attr_setdetachstate(&pattr, PTHREAD_CREATE_DETACHED);
     68   1.3  pooka 	}
     69   1.3  pooka 
     70   1.3  pooka 	rv = pthread_create(ptidp, &pattr, f, arg);
     71   1.9  pooka #if defined(__NetBSD__)
     72   1.1  pooka 	if (rv == 0 && thrname)
     73   1.1  pooka 		pthread_setname_np(ptid, thrname, NULL);
     74   1.9  pooka #elif defined(__linux__)
     75  1.12  pooka 	/*
     76  1.12  pooka 	 * The pthread_setname_np() call varies from one Linux distro to
     77  1.12  pooka 	 * another.  Comment out the call pending autoconf support.
     78  1.12  pooka 	 */
     79  1.12  pooka #if 0
     80   1.9  pooka 	if (rv == 0 && thrname)
     81   1.9  pooka 		pthread_setname_np(ptid, thrname);
     82   1.1  pooka #endif
     83  1.12  pooka #endif
     84   1.1  pooka 
     85   1.3  pooka 	if (joinable) {
     86   1.3  pooka 		assert(ptcookie);
     87   1.3  pooka 		*ptcookie = ptidp;
     88   1.3  pooka 	}
     89   1.3  pooka 
     90   1.3  pooka 	pthread_attr_destroy(&pattr);
     91   1.3  pooka 
     92  1.20  pooka 	ET(rv);
     93   1.1  pooka }
     94   1.1  pooka 
     95   1.1  pooka __dead void
     96   1.1  pooka rumpuser_thread_exit(void)
     97   1.1  pooka {
     98   1.1  pooka 
     99   1.1  pooka 	pthread_exit(NULL);
    100   1.1  pooka }
    101   1.1  pooka 
    102   1.3  pooka int
    103   1.3  pooka rumpuser_thread_join(void *ptcookie)
    104   1.3  pooka {
    105   1.3  pooka 	pthread_t *pt = ptcookie;
    106   1.3  pooka 	int rv;
    107   1.3  pooka 
    108   1.3  pooka 	KLOCK_WRAP((rv = pthread_join(*pt, NULL)));
    109   1.3  pooka 	if (rv == 0)
    110   1.3  pooka 		free(pt);
    111   1.3  pooka 
    112  1.20  pooka 	ET(rv);
    113   1.3  pooka }
    114   1.3  pooka 
    115  1.26  pooka struct rumpuser_mtx {
    116  1.26  pooka 	pthread_mutex_t pthmtx;
    117  1.26  pooka 	struct lwp *owner;
    118  1.26  pooka 	int flags;
    119  1.26  pooka };
    120  1.26  pooka 
    121   1.1  pooka void
    122  1.15  pooka rumpuser_mutex_init(struct rumpuser_mtx **mtx, int flags)
    123   1.1  pooka {
    124   1.1  pooka 	pthread_mutexattr_t att;
    125   1.1  pooka 
    126   1.1  pooka 	NOFAIL(*mtx = malloc(sizeof(struct rumpuser_mtx)));
    127   1.1  pooka 
    128   1.1  pooka 	pthread_mutexattr_init(&att);
    129   1.1  pooka 	pthread_mutexattr_settype(&att, PTHREAD_MUTEX_ERRORCHECK);
    130   1.1  pooka 	NOFAIL_ERRNO(pthread_mutex_init(&((*mtx)->pthmtx), &att));
    131   1.1  pooka 	pthread_mutexattr_destroy(&att);
    132   1.1  pooka 
    133   1.1  pooka 	(*mtx)->owner = NULL;
    134  1.15  pooka 	assert(flags != 0);
    135  1.15  pooka 	(*mtx)->flags = flags;
    136   1.4  pooka }
    137   1.4  pooka 
    138   1.1  pooka static void
    139   1.1  pooka mtxenter(struct rumpuser_mtx *mtx)
    140   1.1  pooka {
    141   1.1  pooka 
    142  1.15  pooka 	if (!(mtx->flags & RUMPUSER_MTX_KMUTEX))
    143   1.4  pooka 		return;
    144   1.4  pooka 
    145   1.4  pooka 	assert(mtx->owner == NULL);
    146  1.23  pooka 	mtx->owner = rumpuser_curlwp();
    147   1.1  pooka }
    148   1.1  pooka 
    149   1.1  pooka static void
    150   1.1  pooka mtxexit(struct rumpuser_mtx *mtx)
    151   1.1  pooka {
    152   1.1  pooka 
    153  1.15  pooka 	if (!(mtx->flags & RUMPUSER_MTX_KMUTEX))
    154   1.4  pooka 		return;
    155   1.4  pooka 
    156   1.1  pooka 	assert(mtx->owner != NULL);
    157   1.4  pooka 	mtx->owner = NULL;
    158   1.1  pooka }
    159   1.1  pooka 
    160   1.1  pooka void
    161   1.1  pooka rumpuser_mutex_enter(struct rumpuser_mtx *mtx)
    162   1.1  pooka {
    163   1.1  pooka 
    164  1.15  pooka 	if (mtx->flags & RUMPUSER_MTX_SPIN) {
    165  1.13  pooka 		rumpuser_mutex_enter_nowrap(mtx);
    166  1.13  pooka 		return;
    167  1.13  pooka 	}
    168  1.13  pooka 
    169  1.15  pooka 	assert(mtx->flags & RUMPUSER_MTX_KMUTEX);
    170   1.1  pooka 	if (pthread_mutex_trylock(&mtx->pthmtx) != 0)
    171   1.1  pooka 		KLOCK_WRAP(NOFAIL_ERRNO(pthread_mutex_lock(&mtx->pthmtx)));
    172   1.1  pooka 	mtxenter(mtx);
    173   1.1  pooka }
    174   1.1  pooka 
    175   1.1  pooka void
    176   1.1  pooka rumpuser_mutex_enter_nowrap(struct rumpuser_mtx *mtx)
    177   1.1  pooka {
    178   1.1  pooka 
    179  1.15  pooka 	assert(mtx->flags & RUMPUSER_MTX_SPIN);
    180   1.1  pooka 	NOFAIL_ERRNO(pthread_mutex_lock(&mtx->pthmtx));
    181   1.1  pooka 	mtxenter(mtx);
    182   1.1  pooka }
    183   1.1  pooka 
    184   1.1  pooka int
    185   1.1  pooka rumpuser_mutex_tryenter(struct rumpuser_mtx *mtx)
    186   1.1  pooka {
    187   1.1  pooka 	int rv;
    188   1.1  pooka 
    189   1.1  pooka 	rv = pthread_mutex_trylock(&mtx->pthmtx);
    190   1.1  pooka 	if (rv == 0) {
    191   1.1  pooka 		mtxenter(mtx);
    192   1.1  pooka 	}
    193   1.1  pooka 
    194  1.20  pooka 	ET(rv);
    195   1.1  pooka }
    196   1.1  pooka 
    197   1.1  pooka void
    198   1.1  pooka rumpuser_mutex_exit(struct rumpuser_mtx *mtx)
    199   1.1  pooka {
    200   1.1  pooka 
    201   1.1  pooka 	mtxexit(mtx);
    202   1.1  pooka 	NOFAIL_ERRNO(pthread_mutex_unlock(&mtx->pthmtx));
    203   1.1  pooka }
    204   1.1  pooka 
    205   1.1  pooka void
    206   1.1  pooka rumpuser_mutex_destroy(struct rumpuser_mtx *mtx)
    207   1.1  pooka {
    208   1.1  pooka 
    209   1.1  pooka 	NOFAIL_ERRNO(pthread_mutex_destroy(&mtx->pthmtx));
    210   1.1  pooka 	free(mtx);
    211   1.1  pooka }
    212   1.1  pooka 
    213  1.19  pooka void
    214  1.19  pooka rumpuser_mutex_owner(struct rumpuser_mtx *mtx, struct lwp **lp)
    215   1.1  pooka {
    216   1.1  pooka 
    217  1.15  pooka 	if (__predict_false(!(mtx->flags & RUMPUSER_MTX_KMUTEX))) {
    218   1.4  pooka 		printf("panic: rumpuser_mutex_held unsupported on non-kmtx\n");
    219   1.4  pooka 		abort();
    220   1.4  pooka 	}
    221   1.4  pooka 
    222  1.19  pooka 	*lp = mtx->owner;
    223   1.1  pooka }
    224   1.1  pooka 
    225  1.26  pooka /*
    226  1.27  pooka  * rwlocks.  these are mostly simple, except that NetBSD wants to
    227  1.27  pooka  * support something called downgrade, which means we need to swap
    228  1.27  pooka  * our exclusive lock for a shared lock.  to accommodate this,
    229  1.27  pooka  * we need to check *after* acquiring a lock in case someone was
    230  1.27  pooka  * downgrading it.  if so, we couldn't actually have it and maybe
    231  1.27  pooka  * need to retry later.
    232  1.26  pooka  */
    233  1.26  pooka 
    234  1.26  pooka struct rumpuser_rw {
    235  1.26  pooka 	pthread_rwlock_t pthrw;
    236  1.26  pooka 	pthread_spinlock_t spin;
    237  1.26  pooka 	int readers;
    238  1.26  pooka 	struct lwp *writer;
    239  1.27  pooka 	int downgrade; /* someone is downgrading (hopefully lock holder ;) */
    240  1.26  pooka };
    241  1.26  pooka 
    242  1.27  pooka static int
    243  1.27  pooka rw_amwriter(struct rumpuser_rw *rw)
    244  1.27  pooka {
    245  1.27  pooka 
    246  1.27  pooka 	return rw->writer == rumpuser_curlwp() && rw->readers == -1;
    247  1.27  pooka }
    248  1.27  pooka 
    249  1.27  pooka static int
    250  1.27  pooka rw_nreaders(struct rumpuser_rw *rw)
    251  1.27  pooka {
    252  1.27  pooka 
    253  1.27  pooka 	return rw->readers > 0 ? rw->readers : 0;
    254  1.27  pooka }
    255  1.27  pooka 
    256  1.27  pooka static int
    257  1.27  pooka rw_setwriter(struct rumpuser_rw *rw, int retry)
    258  1.27  pooka {
    259  1.27  pooka 
    260  1.27  pooka 	/*
    261  1.27  pooka 	 * Don't need the spinlock here, we already have an
    262  1.27  pooka 	 * exclusive lock and "downgrade" is stable until complete.
    263  1.27  pooka 	 */
    264  1.27  pooka 	if (rw->downgrade) {
    265  1.27  pooka 		pthread_rwlock_unlock(&rw->pthrw);
    266  1.27  pooka 		if (retry) {
    267  1.27  pooka 			struct timespec ts;
    268  1.27  pooka 
    269  1.27  pooka 			/* portable yield, essentially */
    270  1.27  pooka 			ts.tv_sec = 0;
    271  1.27  pooka 			ts.tv_nsec = 1;
    272  1.27  pooka 			KLOCK_WRAP(nanosleep(&ts, NULL));
    273  1.27  pooka 		}
    274  1.27  pooka 		return EBUSY;
    275  1.27  pooka 	}
    276  1.27  pooka 	assert(rw->readers == 0);
    277  1.27  pooka 	rw->writer = rumpuser_curlwp();
    278  1.27  pooka 	rw->readers = -1;
    279  1.27  pooka 	return 0;
    280  1.27  pooka }
    281  1.27  pooka 
    282  1.27  pooka static void
    283  1.27  pooka rw_clearwriter(struct rumpuser_rw *rw)
    284  1.27  pooka {
    285  1.27  pooka 
    286  1.27  pooka 	assert(rw_amwriter(rw));
    287  1.27  pooka 	rw->readers = 0;
    288  1.27  pooka 	rw->writer = NULL;
    289  1.27  pooka }
    290  1.27  pooka 
    291  1.27  pooka static void
    292  1.27  pooka rw_readup(struct rumpuser_rw *rw)
    293  1.27  pooka {
    294  1.27  pooka 
    295  1.27  pooka 	pthread_spin_lock(&rw->spin);
    296  1.27  pooka 	assert(rw->readers >= 0);
    297  1.27  pooka 	++rw->readers;
    298  1.27  pooka 	pthread_spin_unlock(&rw->spin);
    299  1.27  pooka }
    300  1.27  pooka 
    301  1.27  pooka static void
    302  1.27  pooka rw_readdown(struct rumpuser_rw *rw)
    303  1.27  pooka {
    304  1.27  pooka 
    305  1.27  pooka 	pthread_spin_lock(&rw->spin);
    306  1.27  pooka 	assert(rw->readers > 0);
    307  1.27  pooka 	--rw->readers;
    308  1.27  pooka 	pthread_spin_unlock(&rw->spin);
    309  1.27  pooka }
    310  1.26  pooka 
    311   1.1  pooka void
    312   1.1  pooka rumpuser_rw_init(struct rumpuser_rw **rw)
    313   1.1  pooka {
    314   1.1  pooka 
    315   1.1  pooka 	NOFAIL(*rw = malloc(sizeof(struct rumpuser_rw)));
    316   1.1  pooka 	NOFAIL_ERRNO(pthread_rwlock_init(&((*rw)->pthrw), NULL));
    317  1.10  pooka 	NOFAIL_ERRNO(pthread_spin_init(&((*rw)->spin),PTHREAD_PROCESS_PRIVATE));
    318   1.1  pooka 	(*rw)->readers = 0;
    319   1.1  pooka 	(*rw)->writer = NULL;
    320  1.28  pooka 	(*rw)->downgrade = 0;
    321   1.1  pooka }
    322   1.1  pooka 
    323   1.1  pooka void
    324  1.25  pooka rumpuser_rw_enter(struct rumpuser_rw *rw, const enum rumprwlock lk)
    325   1.1  pooka {
    326   1.1  pooka 
    327  1.25  pooka 	switch (lk) {
    328  1.25  pooka 	case RUMPUSER_RW_WRITER:
    329  1.27  pooka 		do {
    330  1.27  pooka 			if (pthread_rwlock_trywrlock(&rw->pthrw) != 0)
    331  1.27  pooka 				KLOCK_WRAP(NOFAIL_ERRNO(
    332  1.27  pooka 				    pthread_rwlock_wrlock(&rw->pthrw)));
    333  1.27  pooka 		} while (rw_setwriter(rw, 1) != 0);
    334  1.25  pooka 		break;
    335  1.25  pooka 	case RUMPUSER_RW_READER:
    336   1.1  pooka 		if (pthread_rwlock_tryrdlock(&rw->pthrw) != 0)
    337   1.1  pooka 			KLOCK_WRAP(NOFAIL_ERRNO(
    338   1.1  pooka 			    pthread_rwlock_rdlock(&rw->pthrw)));
    339  1.27  pooka 		rw_readup(rw);
    340  1.25  pooka 		break;
    341   1.1  pooka 	}
    342   1.1  pooka }
    343   1.1  pooka 
    344   1.1  pooka int
    345  1.25  pooka rumpuser_rw_tryenter(struct rumpuser_rw *rw, const enum rumprwlock lk)
    346   1.1  pooka {
    347   1.1  pooka 	int rv;
    348   1.1  pooka 
    349  1.25  pooka 	switch (lk) {
    350  1.25  pooka 	case RUMPUSER_RW_WRITER:
    351   1.1  pooka 		rv = pthread_rwlock_trywrlock(&rw->pthrw);
    352   1.1  pooka 		if (rv == 0)
    353  1.27  pooka 			rv = rw_setwriter(rw, 0);
    354  1.25  pooka 		break;
    355  1.25  pooka 	case RUMPUSER_RW_READER:
    356   1.1  pooka 		rv = pthread_rwlock_tryrdlock(&rw->pthrw);
    357   1.1  pooka 		if (rv == 0)
    358  1.27  pooka 			rw_readup(rw);
    359  1.25  pooka 		break;
    360  1.25  pooka 	default:
    361  1.25  pooka 		rv = EINVAL;
    362  1.25  pooka 		break;
    363   1.1  pooka 	}
    364   1.1  pooka 
    365  1.20  pooka 	ET(rv);
    366   1.1  pooka }
    367   1.1  pooka 
    368  1.25  pooka int
    369  1.25  pooka rumpuser_rw_tryupgrade(struct rumpuser_rw *rw)
    370  1.25  pooka {
    371  1.25  pooka 
    372  1.27  pooka 	/*
    373  1.27  pooka 	 * Not supported by pthreads.  Since the caller needs to
    374  1.27  pooka 	 * back off anyway to avoid deadlock, always failing
    375  1.27  pooka 	 * is correct.
    376  1.27  pooka 	 */
    377  1.25  pooka 	ET(EBUSY);
    378  1.25  pooka }
    379  1.25  pooka 
    380  1.27  pooka /*
    381  1.27  pooka  * convert from exclusive to shared lock without allowing anyone to
    382  1.27  pooka  * obtain an exclusive lock in between.  actually, might allow
    383  1.27  pooka  * someone to obtain the lock, we just don't allow that thread to
    384  1.27  pooka  * return from the hypercall with it.
    385  1.27  pooka  */
    386  1.25  pooka void
    387  1.25  pooka rumpuser_rw_downgrade(struct rumpuser_rw *rw)
    388  1.25  pooka {
    389  1.25  pooka 
    390  1.27  pooka 	assert(rw->downgrade == 0);
    391  1.27  pooka 	rw->downgrade = 1;
    392  1.27  pooka 	rumpuser_rw_exit(rw);
    393  1.25  pooka 	/*
    394  1.27  pooka 	 * though the competition can't get out of the hypervisor, it
    395  1.27  pooka 	 * might have rescheduled itself after we released the lock.
    396  1.27  pooka 	 * so need a wrap here.
    397  1.25  pooka 	 */
    398  1.27  pooka 	KLOCK_WRAP(NOFAIL_ERRNO(pthread_rwlock_rdlock(&rw->pthrw)));
    399  1.27  pooka 	rw->downgrade = 0;
    400  1.27  pooka 	rw_readup(rw);
    401  1.25  pooka }
    402  1.25  pooka 
    403   1.1  pooka void
    404   1.1  pooka rumpuser_rw_exit(struct rumpuser_rw *rw)
    405   1.1  pooka {
    406   1.1  pooka 
    407  1.27  pooka 	if (rw_nreaders(rw))
    408  1.27  pooka 		rw_readdown(rw);
    409   1.1  pooka 	else
    410  1.27  pooka 		rw_clearwriter(rw);
    411   1.1  pooka 	NOFAIL_ERRNO(pthread_rwlock_unlock(&rw->pthrw));
    412   1.1  pooka }
    413   1.1  pooka 
    414   1.1  pooka void
    415   1.1  pooka rumpuser_rw_destroy(struct rumpuser_rw *rw)
    416   1.1  pooka {
    417   1.1  pooka 
    418   1.1  pooka 	NOFAIL_ERRNO(pthread_rwlock_destroy(&rw->pthrw));
    419   1.1  pooka 	NOFAIL_ERRNO(pthread_spin_destroy(&rw->spin));
    420   1.1  pooka 	free(rw);
    421   1.1  pooka }
    422   1.1  pooka 
    423  1.19  pooka void
    424  1.25  pooka rumpuser_rw_held(struct rumpuser_rw *rw, const enum rumprwlock lk, int *rv)
    425   1.1  pooka {
    426   1.1  pooka 
    427  1.25  pooka 	switch (lk) {
    428  1.25  pooka 	case RUMPUSER_RW_WRITER:
    429  1.27  pooka 		*rv = rw_amwriter(rw);
    430  1.25  pooka 		break;
    431  1.25  pooka 	case RUMPUSER_RW_READER:
    432  1.27  pooka 		*rv = rw_nreaders(rw);
    433  1.25  pooka 		break;
    434  1.25  pooka 	}
    435   1.1  pooka }
    436   1.1  pooka 
    437  1.26  pooka /*
    438  1.26  pooka  * condvar
    439  1.26  pooka  */
    440  1.26  pooka 
    441  1.26  pooka struct rumpuser_cv {
    442  1.26  pooka 	pthread_cond_t pthcv;
    443  1.26  pooka 	int nwaiters;
    444  1.26  pooka };
    445  1.26  pooka 
    446   1.1  pooka void
    447   1.1  pooka rumpuser_cv_init(struct rumpuser_cv **cv)
    448   1.1  pooka {
    449   1.1  pooka 
    450   1.1  pooka 	NOFAIL(*cv = malloc(sizeof(struct rumpuser_cv)));
    451   1.1  pooka 	NOFAIL_ERRNO(pthread_cond_init(&((*cv)->pthcv), NULL));
    452   1.1  pooka 	(*cv)->nwaiters = 0;
    453   1.1  pooka }
    454   1.1  pooka 
    455   1.1  pooka void
    456   1.1  pooka rumpuser_cv_destroy(struct rumpuser_cv *cv)
    457   1.1  pooka {
    458   1.1  pooka 
    459   1.1  pooka 	NOFAIL_ERRNO(pthread_cond_destroy(&cv->pthcv));
    460   1.1  pooka 	free(cv);
    461   1.1  pooka }
    462   1.1  pooka 
    463  1.24  pooka static void
    464  1.24  pooka cv_unschedule(struct rumpuser_mtx *mtx, int *nlocks)
    465  1.24  pooka {
    466  1.24  pooka 
    467  1.24  pooka 	rumpkern_unsched(nlocks, mtx);
    468  1.24  pooka 	mtxexit(mtx);
    469  1.24  pooka }
    470  1.24  pooka 
    471  1.24  pooka static void
    472  1.24  pooka cv_reschedule(struct rumpuser_mtx *mtx, int nlocks)
    473  1.24  pooka {
    474  1.24  pooka 
    475  1.24  pooka 	/*
    476  1.24  pooka 	 * If the cv interlock is a spin mutex, we must first release
    477  1.24  pooka 	 * the mutex that was reacquired by pthread_cond_wait(),
    478  1.24  pooka 	 * acquire the CPU context and only then relock the mutex.
    479  1.24  pooka 	 * This is to preserve resource allocation order so that
    480  1.24  pooka 	 * we don't deadlock.  Non-spinning mutexes don't have this
    481  1.24  pooka 	 * problem since they don't use a hold-and-wait approach
    482  1.24  pooka 	 * to acquiring the mutex wrt the rump kernel CPU context.
    483  1.24  pooka 	 *
    484  1.24  pooka 	 * The more optimal solution would be to rework rumpkern_sched()
    485  1.24  pooka 	 * so that it's possible to tell the scheduler
    486  1.24  pooka 	 * "if you need to block, drop this lock first", but I'm not
    487  1.24  pooka 	 * going poking there without some numbers on how often this
    488  1.24  pooka 	 * path is taken for spin mutexes.
    489  1.24  pooka 	 */
    490  1.24  pooka 	if ((mtx->flags & (RUMPUSER_MTX_SPIN | RUMPUSER_MTX_KMUTEX)) ==
    491  1.24  pooka 	    (RUMPUSER_MTX_SPIN | RUMPUSER_MTX_KMUTEX)) {
    492  1.24  pooka 		NOFAIL_ERRNO(pthread_mutex_unlock(&mtx->pthmtx));
    493  1.24  pooka 		rumpkern_sched(nlocks, mtx);
    494  1.24  pooka 		rumpuser_mutex_enter_nowrap(mtx);
    495  1.24  pooka 	} else {
    496  1.24  pooka 		mtxenter(mtx);
    497  1.24  pooka 		rumpkern_sched(nlocks, mtx);
    498  1.24  pooka 	}
    499  1.24  pooka }
    500  1.24  pooka 
    501   1.1  pooka void
    502   1.1  pooka rumpuser_cv_wait(struct rumpuser_cv *cv, struct rumpuser_mtx *mtx)
    503   1.1  pooka {
    504   1.2  pooka 	int nlocks;
    505   1.1  pooka 
    506   1.1  pooka 	cv->nwaiters++;
    507  1.24  pooka 	cv_unschedule(mtx, &nlocks);
    508   1.2  pooka 	NOFAIL_ERRNO(pthread_cond_wait(&cv->pthcv, &mtx->pthmtx));
    509  1.24  pooka 	cv_reschedule(mtx, nlocks);
    510   1.1  pooka 	cv->nwaiters--;
    511   1.1  pooka }
    512   1.1  pooka 
    513   1.1  pooka void
    514   1.1  pooka rumpuser_cv_wait_nowrap(struct rumpuser_cv *cv, struct rumpuser_mtx *mtx)
    515   1.1  pooka {
    516   1.1  pooka 
    517   1.1  pooka 	cv->nwaiters++;
    518   1.1  pooka 	mtxexit(mtx);
    519   1.1  pooka 	NOFAIL_ERRNO(pthread_cond_wait(&cv->pthcv, &mtx->pthmtx));
    520   1.1  pooka 	mtxenter(mtx);
    521   1.1  pooka 	cv->nwaiters--;
    522   1.1  pooka }
    523   1.1  pooka 
    524   1.1  pooka int
    525   1.1  pooka rumpuser_cv_timedwait(struct rumpuser_cv *cv, struct rumpuser_mtx *mtx,
    526   1.1  pooka 	int64_t sec, int64_t nsec)
    527   1.1  pooka {
    528   1.1  pooka 	struct timespec ts;
    529   1.2  pooka 	int rv, nlocks;
    530   1.1  pooka 
    531  1.16  pooka 	/*
    532  1.16  pooka 	 * Get clock already here, just in case we will be put to sleep
    533  1.16  pooka 	 * after releasing the kernel context.
    534  1.16  pooka 	 *
    535  1.16  pooka 	 * The condition variables should use CLOCK_MONOTONIC, but since
    536  1.16  pooka 	 * that's not available everywhere, leave it for another day.
    537  1.16  pooka 	 */
    538  1.16  pooka 	clock_gettime(CLOCK_REALTIME, &ts);
    539   1.1  pooka 
    540   1.1  pooka 	cv->nwaiters++;
    541  1.24  pooka 	cv_unschedule(mtx, &nlocks);
    542  1.16  pooka 
    543  1.16  pooka 	ts.tv_sec += sec;
    544  1.16  pooka 	ts.tv_nsec += nsec;
    545  1.16  pooka 	if (ts.tv_nsec >= 1000*1000*1000) {
    546  1.16  pooka 		ts.tv_sec++;
    547  1.16  pooka 		ts.tv_nsec -= 1000*1000*1000;
    548  1.16  pooka 	}
    549   1.2  pooka 	rv = pthread_cond_timedwait(&cv->pthcv, &mtx->pthmtx, &ts);
    550  1.24  pooka 
    551  1.24  pooka 	cv_reschedule(mtx, nlocks);
    552   1.1  pooka 	cv->nwaiters--;
    553   1.1  pooka 
    554  1.20  pooka 	ET(rv);
    555   1.1  pooka }
    556   1.1  pooka 
    557   1.1  pooka void
    558   1.1  pooka rumpuser_cv_signal(struct rumpuser_cv *cv)
    559   1.1  pooka {
    560   1.1  pooka 
    561   1.1  pooka 	NOFAIL_ERRNO(pthread_cond_signal(&cv->pthcv));
    562   1.1  pooka }
    563   1.1  pooka 
    564   1.1  pooka void
    565   1.1  pooka rumpuser_cv_broadcast(struct rumpuser_cv *cv)
    566   1.1  pooka {
    567   1.1  pooka 
    568   1.1  pooka 	NOFAIL_ERRNO(pthread_cond_broadcast(&cv->pthcv));
    569   1.1  pooka }
    570   1.1  pooka 
    571  1.19  pooka void
    572  1.19  pooka rumpuser_cv_has_waiters(struct rumpuser_cv *cv, int *nwaiters)
    573   1.1  pooka {
    574   1.1  pooka 
    575  1.19  pooka 	*nwaiters = cv->nwaiters;
    576   1.1  pooka }
    577   1.1  pooka 
    578   1.1  pooka /*
    579   1.1  pooka  * curlwp
    580   1.1  pooka  */
    581   1.1  pooka 
    582  1.26  pooka static pthread_key_t curlwpkey;
    583  1.26  pooka 
    584  1.23  pooka /*
    585  1.23  pooka  * the if0'd curlwp implementation is not used by this hypervisor,
    586  1.23  pooka  * but serves as test code to check that the intended usage works.
    587  1.23  pooka  */
    588  1.23  pooka #if 0
    589  1.23  pooka struct rumpuser_lwp {
    590  1.23  pooka 	struct lwp *l;
    591  1.23  pooka 	LIST_ENTRY(rumpuser_lwp) l_entries;
    592  1.23  pooka };
    593  1.23  pooka static LIST_HEAD(, rumpuser_lwp) lwps = LIST_HEAD_INITIALIZER(lwps);
    594  1.23  pooka static pthread_mutex_t lwplock = PTHREAD_MUTEX_INITIALIZER;
    595  1.23  pooka 
    596   1.1  pooka void
    597  1.23  pooka rumpuser_curlwpop(enum rumplwpop op, struct lwp *l)
    598   1.1  pooka {
    599  1.23  pooka 	struct rumpuser_lwp *rl, *rliter;
    600   1.1  pooka 
    601  1.23  pooka 	switch (op) {
    602  1.23  pooka 	case RUMPUSER_LWP_CREATE:
    603  1.23  pooka 		rl = malloc(sizeof(*rl));
    604  1.23  pooka 		rl->l = l;
    605  1.23  pooka 		pthread_mutex_lock(&lwplock);
    606  1.23  pooka 		LIST_FOREACH(rliter, &lwps, l_entries) {
    607  1.23  pooka 			if (rliter->l == l) {
    608  1.23  pooka 				fprintf(stderr, "LWP_CREATE: %p exists\n", l);
    609  1.23  pooka 				abort();
    610  1.23  pooka 			}
    611  1.23  pooka 		}
    612  1.23  pooka 		LIST_INSERT_HEAD(&lwps, rl, l_entries);
    613  1.23  pooka 		pthread_mutex_unlock(&lwplock);
    614  1.23  pooka 		break;
    615  1.23  pooka 	case RUMPUSER_LWP_DESTROY:
    616  1.23  pooka 		pthread_mutex_lock(&lwplock);
    617  1.23  pooka 		LIST_FOREACH(rl, &lwps, l_entries) {
    618  1.23  pooka 			if (rl->l == l)
    619  1.23  pooka 				break;
    620  1.23  pooka 		}
    621  1.23  pooka 		if (!rl) {
    622  1.23  pooka 			fprintf(stderr, "LWP_DESTROY: %p does not exist\n", l);
    623  1.23  pooka 			abort();
    624  1.23  pooka 		}
    625  1.23  pooka 		LIST_REMOVE(rl, l_entries);
    626  1.23  pooka 		pthread_mutex_unlock(&lwplock);
    627  1.23  pooka 		free(rl);
    628  1.23  pooka 		break;
    629  1.23  pooka 	case RUMPUSER_LWP_SET:
    630  1.29  pooka 		assert(pthread_getspecific(curlwpkey) == NULL && l != NULL);
    631  1.23  pooka 
    632  1.29  pooka 		pthread_mutex_lock(&lwplock);
    633  1.29  pooka 		LIST_FOREACH(rl, &lwps, l_entries) {
    634  1.29  pooka 			if (rl->l == l)
    635  1.29  pooka 				break;
    636  1.29  pooka 		}
    637  1.29  pooka 		if (!rl) {
    638  1.29  pooka 			fprintf(stderr,
    639  1.29  pooka 			    "LWP_SET: %p does not exist\n", l);
    640  1.29  pooka 			abort();
    641  1.23  pooka 		}
    642  1.29  pooka 		pthread_mutex_unlock(&lwplock);
    643  1.23  pooka 
    644  1.23  pooka 		pthread_setspecific(curlwpkey, rl);
    645  1.23  pooka 		break;
    646  1.29  pooka 	case RUMPUSER_LWP_CLEAR:
    647  1.29  pooka 		assert(((struct rumpuser_lwp *)
    648  1.29  pooka 		    pthread_getspecific(curlwpkey))->l == l);
    649  1.29  pooka 		pthread_setspecific(curlwpkey, NULL);
    650  1.29  pooka 		break;
    651  1.23  pooka 	}
    652   1.1  pooka }
    653   1.1  pooka 
    654   1.1  pooka struct lwp *
    655  1.23  pooka rumpuser_curlwp(void)
    656  1.23  pooka {
    657  1.23  pooka 	struct rumpuser_lwp *rl;
    658  1.23  pooka 
    659  1.23  pooka 	rl = pthread_getspecific(curlwpkey);
    660  1.23  pooka 	return rl ? rl->l : NULL;
    661  1.23  pooka }
    662  1.23  pooka 
    663  1.23  pooka #else
    664  1.23  pooka 
    665  1.23  pooka void
    666  1.23  pooka rumpuser_curlwpop(enum rumplwpop op, struct lwp *l)
    667  1.23  pooka {
    668  1.23  pooka 
    669  1.23  pooka 	switch (op) {
    670  1.23  pooka 	case RUMPUSER_LWP_CREATE:
    671  1.23  pooka 		break;
    672  1.23  pooka 	case RUMPUSER_LWP_DESTROY:
    673  1.23  pooka 		break;
    674  1.23  pooka 	case RUMPUSER_LWP_SET:
    675  1.29  pooka 		assert(pthread_getspecific(curlwpkey) == NULL);
    676  1.23  pooka 		pthread_setspecific(curlwpkey, l);
    677  1.23  pooka 		break;
    678  1.29  pooka 	case RUMPUSER_LWP_CLEAR:
    679  1.29  pooka 		assert(pthread_getspecific(curlwpkey) == l);
    680  1.29  pooka 		pthread_setspecific(curlwpkey, NULL);
    681  1.29  pooka 		break;
    682  1.23  pooka 	}
    683  1.23  pooka }
    684  1.23  pooka 
    685  1.23  pooka struct lwp *
    686  1.23  pooka rumpuser_curlwp(void)
    687   1.1  pooka {
    688   1.1  pooka 
    689   1.1  pooka 	return pthread_getspecific(curlwpkey);
    690   1.1  pooka }
    691  1.23  pooka #endif
    692  1.26  pooka 
    693  1.26  pooka 
    694  1.26  pooka void
    695  1.26  pooka rumpuser__thrinit(void)
    696  1.26  pooka {
    697  1.26  pooka 	pthread_key_create(&curlwpkey, NULL);
    698  1.26  pooka }
    699