Home | History | Annotate | Line # | Download | only in tprof
tprof.c revision 1.6.2.1
      1  1.6.2.1  uebayasi /*	$NetBSD: tprof.c,v 1.6.2.1 2010/08/17 06:46:42 uebayasi Exp $	*/
      2      1.1      yamt 
      3      1.1      yamt /*-
      4      1.4      yamt  * Copyright (c)2008,2009 YAMAMOTO Takashi,
      5      1.1      yamt  * All rights reserved.
      6      1.1      yamt  *
      7      1.1      yamt  * Redistribution and use in source and binary forms, with or without
      8      1.1      yamt  * modification, are permitted provided that the following conditions
      9      1.1      yamt  * are met:
     10      1.1      yamt  * 1. Redistributions of source code must retain the above copyright
     11      1.1      yamt  *    notice, this list of conditions and the following disclaimer.
     12      1.1      yamt  * 2. Redistributions in binary form must reproduce the above copyright
     13      1.1      yamt  *    notice, this list of conditions and the following disclaimer in the
     14      1.1      yamt  *    documentation and/or other materials provided with the distribution.
     15      1.1      yamt  *
     16      1.1      yamt  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
     17      1.1      yamt  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     18      1.1      yamt  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     19      1.1      yamt  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
     20      1.1      yamt  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     21      1.1      yamt  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     22      1.1      yamt  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     23      1.1      yamt  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     24      1.1      yamt  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     25      1.1      yamt  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     26      1.1      yamt  * SUCH DAMAGE.
     27      1.1      yamt  */
     28      1.1      yamt 
     29      1.1      yamt #include <sys/cdefs.h>
     30  1.6.2.1  uebayasi __KERNEL_RCSID(0, "$NetBSD: tprof.c,v 1.6.2.1 2010/08/17 06:46:42 uebayasi Exp $");
     31      1.1      yamt 
     32      1.1      yamt #include <sys/param.h>
     33      1.1      yamt #include <sys/systm.h>
     34      1.1      yamt #include <sys/kernel.h>
     35      1.1      yamt 
     36      1.1      yamt #include <sys/cpu.h>
     37      1.1      yamt #include <sys/conf.h>
     38      1.1      yamt #include <sys/callout.h>
     39      1.1      yamt #include <sys/kmem.h>
     40      1.4      yamt #include <sys/module.h>
     41      1.1      yamt #include <sys/workqueue.h>
     42      1.1      yamt #include <sys/queue.h>
     43      1.1      yamt 
     44      1.1      yamt #include <dev/tprof/tprof.h>
     45      1.1      yamt #include <dev/tprof/tprof_ioctl.h>
     46      1.1      yamt 
     47      1.4      yamt /*
     48      1.4      yamt  * locking order:
     49      1.4      yamt  *	tprof_reader_lock -> tprof_lock
     50      1.4      yamt  *	tprof_startstop_lock -> tprof_lock
     51      1.4      yamt  */
     52      1.4      yamt 
     53      1.4      yamt /*
     54      1.4      yamt  * protected by:
     55      1.4      yamt  *	L: tprof_lock
     56      1.4      yamt  *	R: tprof_reader_lock
     57      1.4      yamt  *	S: tprof_startstop_lock
     58      1.4      yamt  */
     59      1.4      yamt 
     60      1.1      yamt typedef struct {
     61      1.1      yamt 	uintptr_t s_pc;	/* program counter */
     62      1.1      yamt } tprof_sample_t;
     63      1.1      yamt 
     64      1.1      yamt typedef struct tprof_buf {
     65      1.1      yamt 	u_int b_used;
     66      1.1      yamt 	u_int b_size;
     67      1.1      yamt 	u_int b_overflow;
     68      1.1      yamt 	u_int b_unused;
     69      1.1      yamt 	STAILQ_ENTRY(tprof_buf) b_list;
     70      1.1      yamt 	tprof_sample_t b_data[];
     71      1.1      yamt } tprof_buf_t;
     72      1.1      yamt #define	TPROF_BUF_BYTESIZE(sz) \
     73      1.1      yamt 	(sizeof(tprof_buf_t) + (sz) * sizeof(tprof_sample_t))
     74      1.1      yamt #define	TPROF_MAX_SAMPLES_PER_BUF	10000
     75      1.1      yamt 
     76      1.1      yamt #define	TPROF_MAX_BUF			100
     77      1.1      yamt 
     78      1.1      yamt typedef struct {
     79      1.1      yamt 	tprof_buf_t *c_buf;
     80      1.1      yamt 	struct work c_work;
     81      1.1      yamt 	callout_t c_callout;
     82      1.1      yamt } __aligned(CACHE_LINE_SIZE) tprof_cpu_t;
     83      1.1      yamt 
     84      1.4      yamt typedef struct tprof_backend {
     85      1.4      yamt 	const char *tb_name;
     86      1.4      yamt 	const tprof_backend_ops_t *tb_ops;
     87      1.4      yamt 	LIST_ENTRY(tprof_backend) tb_list;
     88      1.4      yamt 	int tb_usecount;	/* S: */
     89      1.4      yamt } tprof_backend_t;
     90      1.3      yamt 
     91      1.1      yamt static kmutex_t tprof_lock;
     92      1.1      yamt static bool tprof_running;
     93      1.4      yamt static u_int tprof_nworker;		/* L: # of running worker LWPs */
     94      1.1      yamt static lwp_t *tprof_owner;
     95      1.4      yamt static STAILQ_HEAD(, tprof_buf) tprof_list; /* L: global buffer list */
     96      1.4      yamt static u_int tprof_nbuf_on_list;	/* L: # of buffers on tprof_list */
     97      1.1      yamt static struct workqueue *tprof_wq;
     98      1.1      yamt static tprof_cpu_t tprof_cpus[MAXCPUS] __aligned(CACHE_LINE_SIZE);
     99      1.1      yamt static u_int tprof_samples_per_buf;
    100      1.1      yamt 
    101      1.4      yamt static tprof_backend_t *tprof_backend;	/* S: */
    102      1.4      yamt static LIST_HEAD(, tprof_backend) tprof_backends =
    103      1.4      yamt     LIST_HEAD_INITIALIZER(tprof_backend); /* S: */
    104      1.4      yamt 
    105      1.1      yamt static kmutex_t tprof_reader_lock;
    106      1.4      yamt static kcondvar_t tprof_reader_cv;	/* L: */
    107      1.4      yamt static off_t tprof_reader_offset;	/* R: */
    108      1.1      yamt 
    109      1.1      yamt static kmutex_t tprof_startstop_lock;
    110      1.4      yamt static kcondvar_t tprof_cv;		/* L: */
    111      1.1      yamt 
    112      1.4      yamt static struct tprof_stat tprof_stat;	/* L: */
    113      1.1      yamt 
    114      1.1      yamt static tprof_cpu_t *
    115      1.1      yamt tprof_cpu(struct cpu_info *ci)
    116      1.1      yamt {
    117      1.1      yamt 
    118      1.1      yamt 	return &tprof_cpus[cpu_index(ci)];
    119      1.1      yamt }
    120      1.1      yamt 
    121      1.1      yamt static tprof_cpu_t *
    122      1.1      yamt tprof_curcpu(void)
    123      1.1      yamt {
    124      1.1      yamt 
    125      1.1      yamt 	return tprof_cpu(curcpu());
    126      1.1      yamt }
    127      1.1      yamt 
    128      1.1      yamt static tprof_buf_t *
    129      1.1      yamt tprof_buf_alloc(void)
    130      1.1      yamt {
    131      1.1      yamt 	tprof_buf_t *new;
    132      1.1      yamt 	u_int size = tprof_samples_per_buf;
    133      1.1      yamt 
    134      1.1      yamt 	new = kmem_alloc(TPROF_BUF_BYTESIZE(size), KM_SLEEP);
    135      1.1      yamt 	new->b_used = 0;
    136      1.1      yamt 	new->b_size = size;
    137      1.1      yamt 	new->b_overflow = 0;
    138      1.1      yamt 	return new;
    139      1.1      yamt }
    140      1.1      yamt 
    141      1.1      yamt static void
    142      1.1      yamt tprof_buf_free(tprof_buf_t *buf)
    143      1.1      yamt {
    144      1.1      yamt 
    145      1.1      yamt 	kmem_free(buf, TPROF_BUF_BYTESIZE(buf->b_size));
    146      1.1      yamt }
    147      1.1      yamt 
    148      1.1      yamt static tprof_buf_t *
    149      1.1      yamt tprof_buf_switch(tprof_cpu_t *c, tprof_buf_t *new)
    150      1.1      yamt {
    151      1.1      yamt 	tprof_buf_t *old;
    152      1.1      yamt 
    153      1.1      yamt 	old = c->c_buf;
    154      1.1      yamt 	c->c_buf = new;
    155      1.1      yamt 	return old;
    156      1.1      yamt }
    157      1.1      yamt 
    158      1.1      yamt static tprof_buf_t *
    159      1.1      yamt tprof_buf_refresh(void)
    160      1.1      yamt {
    161      1.1      yamt 	tprof_cpu_t * const c = tprof_curcpu();
    162      1.1      yamt 	tprof_buf_t *new;
    163      1.1      yamt 
    164      1.1      yamt 	new = tprof_buf_alloc();
    165      1.1      yamt 	return tprof_buf_switch(c, new);
    166      1.1      yamt }
    167      1.1      yamt 
    168      1.1      yamt static void
    169      1.1      yamt tprof_worker(struct work *wk, void *dummy)
    170      1.1      yamt {
    171      1.1      yamt 	tprof_cpu_t * const c = tprof_curcpu();
    172      1.1      yamt 	tprof_buf_t *buf;
    173      1.1      yamt 	bool shouldstop;
    174      1.1      yamt 
    175      1.1      yamt 	KASSERT(wk == &c->c_work);
    176      1.1      yamt 	KASSERT(dummy == NULL);
    177      1.1      yamt 
    178      1.1      yamt 	/*
    179      1.1      yamt 	 * get a per cpu buffer.
    180      1.1      yamt 	 */
    181      1.1      yamt 	buf = tprof_buf_refresh();
    182      1.1      yamt 
    183      1.1      yamt 	/*
    184      1.1      yamt 	 * and put it on the global list for read(2).
    185      1.1      yamt 	 */
    186      1.1      yamt 	mutex_enter(&tprof_lock);
    187      1.1      yamt 	shouldstop = !tprof_running;
    188      1.1      yamt 	if (shouldstop) {
    189      1.1      yamt 		KASSERT(tprof_nworker > 0);
    190      1.1      yamt 		tprof_nworker--;
    191      1.1      yamt 		cv_broadcast(&tprof_cv);
    192      1.1      yamt 		cv_broadcast(&tprof_reader_cv);
    193      1.1      yamt 	}
    194      1.1      yamt 	if (buf->b_used == 0) {
    195      1.1      yamt 		tprof_stat.ts_emptybuf++;
    196      1.1      yamt 	} else if (tprof_nbuf_on_list < TPROF_MAX_BUF) {
    197      1.1      yamt 		tprof_stat.ts_sample += buf->b_used;
    198      1.1      yamt 		tprof_stat.ts_overflow += buf->b_overflow;
    199      1.1      yamt 		tprof_stat.ts_buf++;
    200      1.1      yamt 		STAILQ_INSERT_TAIL(&tprof_list, buf, b_list);
    201      1.1      yamt 		tprof_nbuf_on_list++;
    202      1.1      yamt 		buf = NULL;
    203      1.1      yamt 		cv_broadcast(&tprof_reader_cv);
    204      1.1      yamt 	} else {
    205      1.1      yamt 		tprof_stat.ts_dropbuf_sample += buf->b_used;
    206      1.1      yamt 		tprof_stat.ts_dropbuf++;
    207      1.1      yamt 	}
    208      1.1      yamt 	mutex_exit(&tprof_lock);
    209      1.1      yamt 	if (buf) {
    210      1.1      yamt 		tprof_buf_free(buf);
    211      1.1      yamt 	}
    212      1.1      yamt 	if (!shouldstop) {
    213      1.1      yamt 		callout_schedule(&c->c_callout, hz);
    214      1.1      yamt 	}
    215      1.1      yamt }
    216      1.1      yamt 
    217      1.1      yamt static void
    218      1.1      yamt tprof_kick(void *vp)
    219      1.1      yamt {
    220      1.1      yamt 	struct cpu_info * const ci = vp;
    221      1.1      yamt 	tprof_cpu_t * const c = tprof_cpu(ci);
    222      1.1      yamt 
    223      1.1      yamt 	workqueue_enqueue(tprof_wq, &c->c_work, ci);
    224      1.1      yamt }
    225      1.1      yamt 
    226      1.1      yamt static void
    227      1.1      yamt tprof_stop1(void)
    228      1.1      yamt {
    229      1.1      yamt 	CPU_INFO_ITERATOR cii;
    230      1.1      yamt 	struct cpu_info *ci;
    231      1.1      yamt 
    232      1.1      yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    233      1.6      yamt 	KASSERT(tprof_nworker == 0);
    234      1.1      yamt 
    235      1.1      yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    236      1.1      yamt 		tprof_cpu_t * const c = tprof_cpu(ci);
    237      1.1      yamt 		tprof_buf_t *old;
    238      1.1      yamt 
    239      1.1      yamt 		old = tprof_buf_switch(c, NULL);
    240      1.1      yamt 		if (old != NULL) {
    241      1.1      yamt 			tprof_buf_free(old);
    242      1.1      yamt 		}
    243      1.1      yamt 		callout_destroy(&c->c_callout);
    244      1.1      yamt 	}
    245      1.1      yamt 	workqueue_destroy(tprof_wq);
    246      1.1      yamt }
    247      1.1      yamt 
    248      1.1      yamt static int
    249      1.1      yamt tprof_start(const struct tprof_param *param)
    250      1.1      yamt {
    251      1.1      yamt 	CPU_INFO_ITERATOR cii;
    252      1.1      yamt 	struct cpu_info *ci;
    253      1.1      yamt 	int error;
    254      1.1      yamt 	uint64_t freq;
    255      1.4      yamt 	tprof_backend_t *tb;
    256      1.1      yamt 
    257      1.1      yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    258      1.1      yamt 	if (tprof_running) {
    259      1.1      yamt 		error = EBUSY;
    260      1.1      yamt 		goto done;
    261      1.1      yamt 	}
    262      1.1      yamt 
    263      1.4      yamt 	tb = tprof_backend;
    264      1.4      yamt 	if (tb == NULL) {
    265      1.4      yamt 		error = ENOENT;
    266      1.4      yamt 		goto done;
    267      1.4      yamt 	}
    268      1.4      yamt 	if (tb->tb_usecount > 0) {
    269      1.4      yamt 		error = EBUSY;
    270      1.4      yamt 		goto done;
    271      1.4      yamt 	}
    272      1.4      yamt 
    273      1.4      yamt 	tb->tb_usecount++;
    274      1.4      yamt 	freq = tb->tb_ops->tbo_estimate_freq();
    275      1.1      yamt 	tprof_samples_per_buf = MIN(freq * 2, TPROF_MAX_SAMPLES_PER_BUF);
    276      1.1      yamt 
    277      1.1      yamt 	error = workqueue_create(&tprof_wq, "tprofmv", tprof_worker, NULL,
    278      1.2      yamt 	    PRI_NONE, IPL_SOFTCLOCK, WQ_MPSAFE | WQ_PERCPU);
    279      1.1      yamt 	if (error != 0) {
    280      1.1      yamt 		goto done;
    281      1.1      yamt 	}
    282      1.1      yamt 
    283      1.1      yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    284      1.1      yamt 		tprof_cpu_t * const c = tprof_cpu(ci);
    285      1.1      yamt 		tprof_buf_t *new;
    286      1.1      yamt 		tprof_buf_t *old;
    287      1.1      yamt 
    288      1.1      yamt 		new = tprof_buf_alloc();
    289      1.1      yamt 		old = tprof_buf_switch(c, new);
    290      1.1      yamt 		if (old != NULL) {
    291      1.1      yamt 			tprof_buf_free(old);
    292      1.1      yamt 		}
    293      1.1      yamt 		callout_init(&c->c_callout, CALLOUT_MPSAFE);
    294      1.1      yamt 		callout_setfunc(&c->c_callout, tprof_kick, ci);
    295      1.1      yamt 	}
    296      1.1      yamt 
    297      1.4      yamt 	error = tb->tb_ops->tbo_start(NULL);
    298      1.1      yamt 	if (error != 0) {
    299      1.1      yamt 		tprof_stop1();
    300      1.1      yamt 		goto done;
    301      1.1      yamt 	}
    302      1.1      yamt 
    303      1.1      yamt 	mutex_enter(&tprof_lock);
    304      1.1      yamt 	tprof_running = true;
    305      1.1      yamt 	mutex_exit(&tprof_lock);
    306      1.1      yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    307      1.1      yamt 		tprof_cpu_t * const c = tprof_cpu(ci);
    308      1.1      yamt 
    309      1.1      yamt 		mutex_enter(&tprof_lock);
    310      1.1      yamt 		tprof_nworker++;
    311      1.1      yamt 		mutex_exit(&tprof_lock);
    312      1.1      yamt 		workqueue_enqueue(tprof_wq, &c->c_work, ci);
    313      1.1      yamt 	}
    314      1.1      yamt done:
    315      1.1      yamt 	return error;
    316      1.1      yamt }
    317      1.1      yamt 
    318      1.1      yamt static void
    319      1.1      yamt tprof_stop(void)
    320      1.1      yamt {
    321      1.1      yamt 	CPU_INFO_ITERATOR cii;
    322      1.1      yamt 	struct cpu_info *ci;
    323      1.4      yamt 	tprof_backend_t *tb;
    324      1.1      yamt 
    325      1.1      yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    326      1.1      yamt 	if (!tprof_running) {
    327      1.1      yamt 		goto done;
    328      1.1      yamt 	}
    329      1.1      yamt 
    330      1.4      yamt 	tb = tprof_backend;
    331      1.4      yamt 	KASSERT(tb->tb_usecount > 0);
    332      1.4      yamt 	tb->tb_ops->tbo_stop(NULL);
    333      1.4      yamt 	tb->tb_usecount--;
    334      1.1      yamt 
    335      1.1      yamt 	mutex_enter(&tprof_lock);
    336      1.1      yamt 	tprof_running = false;
    337      1.1      yamt 	cv_broadcast(&tprof_reader_cv);
    338      1.1      yamt 	mutex_exit(&tprof_lock);
    339      1.1      yamt 
    340      1.1      yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    341      1.1      yamt 		mutex_enter(&tprof_lock);
    342      1.1      yamt 		while (tprof_nworker > 0) {
    343      1.1      yamt 			cv_wait(&tprof_cv, &tprof_lock);
    344      1.1      yamt 		}
    345      1.1      yamt 		mutex_exit(&tprof_lock);
    346      1.1      yamt 	}
    347      1.1      yamt 
    348      1.1      yamt 	tprof_stop1();
    349      1.1      yamt done:
    350      1.1      yamt 	;
    351      1.1      yamt }
    352      1.1      yamt 
    353      1.4      yamt /*
    354      1.4      yamt  * tprof_clear: drain unread samples.
    355      1.4      yamt  */
    356      1.4      yamt 
    357      1.1      yamt static void
    358      1.1      yamt tprof_clear(void)
    359      1.1      yamt {
    360      1.1      yamt 	tprof_buf_t *buf;
    361      1.1      yamt 
    362      1.1      yamt 	mutex_enter(&tprof_reader_lock);
    363      1.1      yamt 	mutex_enter(&tprof_lock);
    364      1.1      yamt 	while ((buf = STAILQ_FIRST(&tprof_list)) != NULL) {
    365      1.1      yamt 		if (buf != NULL) {
    366      1.1      yamt 			STAILQ_REMOVE_HEAD(&tprof_list, b_list);
    367      1.1      yamt 			KASSERT(tprof_nbuf_on_list > 0);
    368      1.1      yamt 			tprof_nbuf_on_list--;
    369      1.1      yamt 			mutex_exit(&tprof_lock);
    370      1.1      yamt 			tprof_buf_free(buf);
    371      1.1      yamt 			mutex_enter(&tprof_lock);
    372      1.1      yamt 		}
    373      1.1      yamt 	}
    374      1.1      yamt 	KASSERT(tprof_nbuf_on_list == 0);
    375      1.1      yamt 	mutex_exit(&tprof_lock);
    376      1.1      yamt 	tprof_reader_offset = 0;
    377      1.1      yamt 	mutex_exit(&tprof_reader_lock);
    378      1.1      yamt 
    379      1.1      yamt 	memset(&tprof_stat, 0, sizeof(tprof_stat));
    380      1.1      yamt }
    381      1.1      yamt 
    382      1.4      yamt static tprof_backend_t *
    383      1.4      yamt tprof_backend_lookup(const char *name)
    384      1.4      yamt {
    385      1.4      yamt 	tprof_backend_t *tb;
    386      1.4      yamt 
    387      1.4      yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    388      1.4      yamt 
    389      1.4      yamt 	LIST_FOREACH(tb, &tprof_backends, tb_list) {
    390      1.4      yamt 		if (!strcmp(tb->tb_name, name)) {
    391      1.4      yamt 			return tb;
    392      1.4      yamt 		}
    393      1.4      yamt 	}
    394      1.4      yamt 	return NULL;
    395      1.4      yamt }
    396      1.4      yamt 
    397      1.1      yamt /* -------------------- backend interfaces */
    398      1.1      yamt 
    399      1.1      yamt /*
    400      1.1      yamt  * tprof_sample: record a sample on the per-cpu buffer.
    401      1.1      yamt  *
    402      1.1      yamt  * be careful; can be called in NMI context.
    403      1.1      yamt  * we are assuming that curcpu() is safe.
    404      1.1      yamt  */
    405      1.1      yamt 
    406      1.1      yamt void
    407      1.5      yamt tprof_sample(tprof_backend_cookie_t *cookie, const tprof_frame_info_t *tfi)
    408      1.1      yamt {
    409      1.1      yamt 	tprof_cpu_t * const c = tprof_curcpu();
    410      1.1      yamt 	tprof_buf_t * const buf = c->c_buf;
    411      1.5      yamt 	const uintptr_t pc = tfi->tfi_pc;
    412      1.1      yamt 	u_int idx;
    413      1.1      yamt 
    414      1.1      yamt 	idx = buf->b_used;
    415      1.1      yamt 	if (__predict_false(idx >= buf->b_size)) {
    416      1.1      yamt 		buf->b_overflow++;
    417      1.1      yamt 		return;
    418      1.1      yamt 	}
    419      1.1      yamt 	buf->b_data[idx].s_pc = pc;
    420      1.1      yamt 	buf->b_used = idx + 1;
    421      1.1      yamt }
    422      1.1      yamt 
    423      1.4      yamt /*
    424      1.4      yamt  * tprof_backend_register:
    425      1.4      yamt  */
    426      1.4      yamt 
    427      1.4      yamt int
    428      1.4      yamt tprof_backend_register(const char *name, const tprof_backend_ops_t *ops,
    429      1.4      yamt     int vers)
    430      1.4      yamt {
    431      1.4      yamt 	tprof_backend_t *tb;
    432      1.4      yamt 
    433      1.4      yamt 	if (vers != TPROF_BACKEND_VERSION) {
    434      1.4      yamt 		return EINVAL;
    435      1.4      yamt 	}
    436      1.4      yamt 
    437      1.4      yamt 	mutex_enter(&tprof_startstop_lock);
    438      1.4      yamt 	tb = tprof_backend_lookup(name);
    439      1.4      yamt 	if (tb != NULL) {
    440      1.4      yamt 		mutex_exit(&tprof_startstop_lock);
    441      1.4      yamt 		return EEXIST;
    442      1.4      yamt 	}
    443      1.4      yamt #if 1 /* XXX for now */
    444      1.4      yamt 	if (!LIST_EMPTY(&tprof_backends)) {
    445      1.4      yamt 		mutex_exit(&tprof_startstop_lock);
    446      1.4      yamt 		return ENOTSUP;
    447      1.4      yamt 	}
    448      1.4      yamt #endif
    449      1.4      yamt 	tb = kmem_alloc(sizeof(*tb), KM_SLEEP);
    450      1.4      yamt 	tb->tb_name = name;
    451      1.4      yamt 	tb->tb_ops = ops;
    452      1.4      yamt 	tb->tb_usecount = 0;
    453      1.4      yamt 	LIST_INSERT_HEAD(&tprof_backends, tb, tb_list);
    454      1.4      yamt #if 1 /* XXX for now */
    455      1.4      yamt 	if (tprof_backend == NULL) {
    456      1.4      yamt 		tprof_backend = tb;
    457      1.4      yamt 	}
    458      1.4      yamt #endif
    459      1.4      yamt 	mutex_exit(&tprof_startstop_lock);
    460      1.4      yamt 
    461      1.4      yamt 	return 0;
    462      1.4      yamt }
    463      1.4      yamt 
    464      1.4      yamt /*
    465      1.4      yamt  * tprof_backend_unregister:
    466      1.4      yamt  */
    467      1.4      yamt 
    468      1.4      yamt int
    469      1.4      yamt tprof_backend_unregister(const char *name)
    470      1.4      yamt {
    471      1.4      yamt 	tprof_backend_t *tb;
    472      1.4      yamt 
    473      1.4      yamt 	mutex_enter(&tprof_startstop_lock);
    474      1.4      yamt 	tb = tprof_backend_lookup(name);
    475      1.4      yamt #if defined(DIAGNOSTIC)
    476      1.4      yamt 	if (tb == NULL) {
    477      1.4      yamt 		mutex_exit(&tprof_startstop_lock);
    478      1.4      yamt 		panic("%s: not found '%s'", __func__, name);
    479      1.4      yamt 	}
    480      1.4      yamt #endif /* defined(DIAGNOSTIC) */
    481      1.4      yamt 	if (tb->tb_usecount > 0) {
    482      1.4      yamt 		mutex_exit(&tprof_startstop_lock);
    483      1.4      yamt 		return EBUSY;
    484      1.4      yamt 	}
    485      1.4      yamt #if 1 /* XXX for now */
    486      1.4      yamt 	if (tprof_backend == tb) {
    487      1.4      yamt 		tprof_backend = NULL;
    488      1.4      yamt 	}
    489      1.4      yamt #endif
    490      1.4      yamt 	LIST_REMOVE(tb, tb_list);
    491      1.4      yamt 	mutex_exit(&tprof_startstop_lock);
    492      1.4      yamt 
    493      1.4      yamt 	kmem_free(tb, sizeof(*tb));
    494      1.4      yamt 
    495      1.4      yamt 	return 0;
    496      1.4      yamt }
    497      1.4      yamt 
    498      1.1      yamt /* -------------------- cdevsw interfaces */
    499      1.1      yamt 
    500      1.1      yamt void tprofattach(int);
    501      1.1      yamt 
    502      1.1      yamt static int
    503      1.1      yamt tprof_open(dev_t dev, int flags, int type, struct lwp *l)
    504      1.1      yamt {
    505      1.1      yamt 
    506      1.1      yamt 	if (minor(dev) != 0) {
    507      1.1      yamt 		return EXDEV;
    508      1.1      yamt 	}
    509      1.1      yamt 	mutex_enter(&tprof_lock);
    510      1.1      yamt 	if (tprof_owner != NULL) {
    511      1.1      yamt 		mutex_exit(&tprof_lock);
    512      1.1      yamt 		return  EBUSY;
    513      1.1      yamt 	}
    514      1.1      yamt 	tprof_owner = curlwp;
    515      1.1      yamt 	mutex_exit(&tprof_lock);
    516      1.1      yamt 
    517      1.1      yamt 	return 0;
    518      1.1      yamt }
    519      1.1      yamt 
    520      1.1      yamt static int
    521      1.1      yamt tprof_close(dev_t dev, int flags, int type, struct lwp *l)
    522      1.1      yamt {
    523      1.1      yamt 
    524      1.1      yamt 	KASSERT(minor(dev) == 0);
    525      1.1      yamt 
    526      1.1      yamt 	mutex_enter(&tprof_startstop_lock);
    527      1.1      yamt 	mutex_enter(&tprof_lock);
    528      1.1      yamt 	tprof_owner = NULL;
    529      1.1      yamt 	mutex_exit(&tprof_lock);
    530      1.1      yamt 	tprof_stop();
    531      1.1      yamt 	tprof_clear();
    532      1.1      yamt 	mutex_exit(&tprof_startstop_lock);
    533      1.1      yamt 
    534      1.1      yamt 	return 0;
    535      1.1      yamt }
    536      1.1      yamt 
    537      1.1      yamt static int
    538      1.1      yamt tprof_read(dev_t dev, struct uio *uio, int flags)
    539      1.1      yamt {
    540      1.1      yamt 	tprof_buf_t *buf;
    541      1.1      yamt 	size_t bytes;
    542      1.1      yamt 	size_t resid;
    543      1.1      yamt 	size_t done;
    544      1.1      yamt 	int error = 0;
    545      1.1      yamt 
    546      1.1      yamt 	KASSERT(minor(dev) == 0);
    547      1.1      yamt 	mutex_enter(&tprof_reader_lock);
    548      1.1      yamt 	while (uio->uio_resid > 0 && error == 0) {
    549      1.1      yamt 		/*
    550      1.1      yamt 		 * take the first buffer from the list.
    551      1.1      yamt 		 */
    552      1.1      yamt 		mutex_enter(&tprof_lock);
    553      1.1      yamt 		buf = STAILQ_FIRST(&tprof_list);
    554      1.1      yamt 		if (buf == NULL) {
    555      1.1      yamt 			if (tprof_nworker == 0) {
    556      1.1      yamt 				mutex_exit(&tprof_lock);
    557      1.1      yamt 				error = 0;
    558      1.1      yamt 				break;
    559      1.1      yamt 			}
    560      1.1      yamt 			mutex_exit(&tprof_reader_lock);
    561      1.1      yamt 			error = cv_wait_sig(&tprof_reader_cv, &tprof_lock);
    562      1.1      yamt 			mutex_exit(&tprof_lock);
    563      1.1      yamt 			mutex_enter(&tprof_reader_lock);
    564      1.1      yamt 			continue;
    565      1.1      yamt 		}
    566      1.1      yamt 		STAILQ_REMOVE_HEAD(&tprof_list, b_list);
    567      1.1      yamt 		KASSERT(tprof_nbuf_on_list > 0);
    568      1.1      yamt 		tprof_nbuf_on_list--;
    569      1.1      yamt 		mutex_exit(&tprof_lock);
    570      1.1      yamt 
    571      1.1      yamt 		/*
    572      1.1      yamt 		 * copy it out.
    573      1.1      yamt 		 */
    574      1.1      yamt 		bytes = MIN(buf->b_used * sizeof(tprof_sample_t) -
    575      1.1      yamt 		    tprof_reader_offset, uio->uio_resid);
    576      1.1      yamt 		resid = uio->uio_resid;
    577      1.1      yamt 		error = uiomove((char *)buf->b_data + tprof_reader_offset,
    578      1.1      yamt 		    bytes, uio);
    579      1.1      yamt 		done = resid - uio->uio_resid;
    580      1.1      yamt 		tprof_reader_offset += done;
    581      1.1      yamt 
    582      1.1      yamt 		/*
    583      1.1      yamt 		 * if we didn't consume the whole buffer,
    584      1.1      yamt 		 * put it back to the list.
    585      1.1      yamt 		 */
    586      1.1      yamt 		if (tprof_reader_offset <
    587      1.1      yamt 		    buf->b_used * sizeof(tprof_sample_t)) {
    588      1.1      yamt 			mutex_enter(&tprof_lock);
    589      1.1      yamt 			STAILQ_INSERT_HEAD(&tprof_list, buf, b_list);
    590      1.1      yamt 			tprof_nbuf_on_list++;
    591      1.1      yamt 			cv_broadcast(&tprof_reader_cv);
    592      1.1      yamt 			mutex_exit(&tprof_lock);
    593      1.1      yamt 		} else {
    594      1.1      yamt 			tprof_buf_free(buf);
    595      1.1      yamt 			tprof_reader_offset = 0;
    596      1.1      yamt 		}
    597      1.1      yamt 	}
    598      1.1      yamt 	mutex_exit(&tprof_reader_lock);
    599      1.1      yamt 
    600      1.1      yamt 	return error;
    601      1.1      yamt }
    602      1.1      yamt 
    603      1.1      yamt static int
    604      1.1      yamt tprof_ioctl(dev_t dev, u_long cmd, void *data, int flags, struct lwp *l)
    605      1.1      yamt {
    606      1.1      yamt 	const struct tprof_param *param;
    607      1.1      yamt 	int error = 0;
    608      1.1      yamt 
    609      1.1      yamt 	KASSERT(minor(dev) == 0);
    610      1.1      yamt 
    611      1.1      yamt 	switch (cmd) {
    612      1.1      yamt 	case TPROF_IOC_GETVERSION:
    613      1.1      yamt 		*(int *)data = TPROF_VERSION;
    614      1.1      yamt 		break;
    615      1.1      yamt 	case TPROF_IOC_START:
    616      1.1      yamt 		param = data;
    617      1.1      yamt 		mutex_enter(&tprof_startstop_lock);
    618      1.1      yamt 		error = tprof_start(param);
    619      1.1      yamt 		mutex_exit(&tprof_startstop_lock);
    620      1.1      yamt 		break;
    621      1.1      yamt 	case TPROF_IOC_STOP:
    622      1.1      yamt 		mutex_enter(&tprof_startstop_lock);
    623      1.1      yamt 		tprof_stop();
    624      1.1      yamt 		mutex_exit(&tprof_startstop_lock);
    625      1.1      yamt 		break;
    626      1.1      yamt 	case TPROF_IOC_GETSTAT:
    627      1.1      yamt 		mutex_enter(&tprof_lock);
    628      1.1      yamt 		memcpy(data, &tprof_stat, sizeof(tprof_stat));
    629      1.1      yamt 		mutex_exit(&tprof_lock);
    630      1.1      yamt 		break;
    631      1.1      yamt 	default:
    632      1.1      yamt 		error = EINVAL;
    633      1.1      yamt 		break;
    634      1.1      yamt 	}
    635      1.1      yamt 
    636      1.1      yamt 	return error;
    637      1.1      yamt }
    638      1.1      yamt 
    639      1.1      yamt const struct cdevsw tprof_cdevsw = {
    640      1.1      yamt 	.d_open = tprof_open,
    641      1.1      yamt 	.d_close = tprof_close,
    642      1.1      yamt 	.d_read = tprof_read,
    643      1.1      yamt 	.d_write = nowrite,
    644      1.1      yamt 	.d_ioctl = tprof_ioctl,
    645      1.1      yamt 	.d_stop = nostop,
    646      1.1      yamt 	.d_tty = notty,
    647      1.1      yamt 	.d_poll = nopoll,
    648      1.1      yamt 	.d_mmap = nommap,
    649      1.1      yamt 	.d_kqfilter = nokqfilter,
    650      1.1      yamt 	.d_flag = D_OTHER | D_MPSAFE,
    651      1.1      yamt };
    652      1.1      yamt 
    653      1.1      yamt void
    654      1.1      yamt tprofattach(int nunits)
    655      1.1      yamt {
    656      1.1      yamt 
    657      1.4      yamt 	/* nothing */
    658      1.4      yamt }
    659      1.4      yamt 
    660      1.4      yamt MODULE(MODULE_CLASS_DRIVER, tprof, NULL);
    661      1.4      yamt 
    662      1.4      yamt static void
    663      1.4      yamt tprof_driver_init(void)
    664      1.4      yamt {
    665      1.4      yamt 
    666      1.1      yamt 	mutex_init(&tprof_lock, MUTEX_DEFAULT, IPL_NONE);
    667      1.1      yamt 	mutex_init(&tprof_reader_lock, MUTEX_DEFAULT, IPL_NONE);
    668      1.1      yamt 	mutex_init(&tprof_startstop_lock, MUTEX_DEFAULT, IPL_NONE);
    669      1.1      yamt 	cv_init(&tprof_cv, "tprof");
    670  1.6.2.1  uebayasi 	cv_init(&tprof_reader_cv, "tprof_rd");
    671      1.1      yamt 	STAILQ_INIT(&tprof_list);
    672      1.1      yamt }
    673      1.4      yamt 
    674      1.4      yamt static void
    675      1.4      yamt tprof_driver_fini(void)
    676      1.4      yamt {
    677      1.4      yamt 
    678      1.4      yamt 	mutex_destroy(&tprof_lock);
    679      1.4      yamt 	mutex_destroy(&tprof_reader_lock);
    680      1.4      yamt 	mutex_destroy(&tprof_startstop_lock);
    681      1.4      yamt 	cv_destroy(&tprof_cv);
    682      1.4      yamt 	cv_destroy(&tprof_reader_cv);
    683      1.4      yamt }
    684      1.4      yamt 
    685      1.4      yamt static int
    686      1.4      yamt tprof_modcmd(modcmd_t cmd, void *arg)
    687      1.4      yamt {
    688      1.4      yamt 
    689      1.4      yamt 	switch (cmd) {
    690      1.4      yamt 	case MODULE_CMD_INIT:
    691      1.4      yamt 		tprof_driver_init();
    692      1.4      yamt #if defined(_MODULE)
    693      1.4      yamt 		{
    694      1.4      yamt 			devmajor_t bmajor = NODEVMAJOR;
    695      1.4      yamt 			devmajor_t cmajor = NODEVMAJOR;
    696      1.4      yamt 			int error;
    697      1.4      yamt 
    698      1.4      yamt 			error = devsw_attach("tprof", NULL, &bmajor,
    699      1.4      yamt 			    &tprof_cdevsw, &cmajor);
    700      1.4      yamt 			if (error) {
    701      1.4      yamt 				tprof_driver_fini();
    702      1.4      yamt 				return error;
    703      1.4      yamt 			}
    704      1.4      yamt 		}
    705      1.4      yamt #endif /* defined(_MODULE) */
    706      1.4      yamt 		return 0;
    707      1.4      yamt 
    708      1.4      yamt 	case MODULE_CMD_FINI:
    709      1.4      yamt #if defined(_MODULE)
    710      1.4      yamt 		{
    711      1.4      yamt 			int error;
    712      1.4      yamt 			error = devsw_detach(NULL, &tprof_cdevsw);
    713      1.4      yamt 			if (error) {
    714      1.4      yamt 				return error;
    715      1.4      yamt 			}
    716      1.4      yamt 		}
    717      1.4      yamt #endif /* defined(_MODULE) */
    718      1.4      yamt 		tprof_driver_fini();
    719      1.4      yamt 		return 0;
    720      1.4      yamt 
    721      1.4      yamt 	default:
    722      1.4      yamt 		return ENOTTY;
    723      1.4      yamt 	}
    724      1.4      yamt }
    725