Home | History | Annotate | Line # | Download | only in tprof
tprof.c revision 1.3.2.1
      1  1.3.2.1   jym /*	$NetBSD: tprof.c,v 1.3.2.1 2009/05/13 17:21:34 jym Exp $	*/
      2      1.1  yamt 
      3      1.1  yamt /*-
      4  1.3.2.1   jym  * Copyright (c)2008,2009 YAMAMOTO Takashi,
      5      1.1  yamt  * All rights reserved.
      6      1.1  yamt  *
      7      1.1  yamt  * Redistribution and use in source and binary forms, with or without
      8      1.1  yamt  * modification, are permitted provided that the following conditions
      9      1.1  yamt  * are met:
     10      1.1  yamt  * 1. Redistributions of source code must retain the above copyright
     11      1.1  yamt  *    notice, this list of conditions and the following disclaimer.
     12      1.1  yamt  * 2. Redistributions in binary form must reproduce the above copyright
     13      1.1  yamt  *    notice, this list of conditions and the following disclaimer in the
     14      1.1  yamt  *    documentation and/or other materials provided with the distribution.
     15      1.1  yamt  *
     16      1.1  yamt  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
     17      1.1  yamt  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
     18      1.1  yamt  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
     19      1.1  yamt  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
     20      1.1  yamt  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
     21      1.1  yamt  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
     22      1.1  yamt  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
     23      1.1  yamt  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
     24      1.1  yamt  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     25      1.1  yamt  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     26      1.1  yamt  * SUCH DAMAGE.
     27      1.1  yamt  */
     28      1.1  yamt 
     29      1.1  yamt #include <sys/cdefs.h>
     30  1.3.2.1   jym __KERNEL_RCSID(0, "$NetBSD: tprof.c,v 1.3.2.1 2009/05/13 17:21:34 jym Exp $");
     31      1.1  yamt 
     32      1.1  yamt #include <sys/param.h>
     33      1.1  yamt #include <sys/systm.h>
     34      1.1  yamt #include <sys/kernel.h>
     35      1.1  yamt 
     36      1.1  yamt #include <sys/cpu.h>
     37      1.1  yamt #include <sys/conf.h>
     38      1.1  yamt #include <sys/callout.h>
     39      1.1  yamt #include <sys/kmem.h>
     40  1.3.2.1   jym #include <sys/module.h>
     41      1.1  yamt #include <sys/workqueue.h>
     42      1.1  yamt #include <sys/queue.h>
     43      1.1  yamt 
     44      1.1  yamt #include <dev/tprof/tprof.h>
     45      1.1  yamt #include <dev/tprof/tprof_ioctl.h>
     46      1.1  yamt 
     47  1.3.2.1   jym /*
     48  1.3.2.1   jym  * locking order:
     49  1.3.2.1   jym  *	tprof_reader_lock -> tprof_lock
     50  1.3.2.1   jym  *	tprof_startstop_lock -> tprof_lock
     51  1.3.2.1   jym  */
     52  1.3.2.1   jym 
     53  1.3.2.1   jym /*
     54  1.3.2.1   jym  * protected by:
     55  1.3.2.1   jym  *	L: tprof_lock
     56  1.3.2.1   jym  *	R: tprof_reader_lock
     57  1.3.2.1   jym  *	S: tprof_startstop_lock
     58  1.3.2.1   jym  */
     59      1.1  yamt 
     60      1.1  yamt typedef struct {
     61      1.1  yamt 	uintptr_t s_pc;	/* program counter */
     62      1.1  yamt } tprof_sample_t;
     63      1.1  yamt 
     64      1.1  yamt typedef struct tprof_buf {
     65      1.1  yamt 	u_int b_used;
     66      1.1  yamt 	u_int b_size;
     67      1.1  yamt 	u_int b_overflow;
     68      1.1  yamt 	u_int b_unused;
     69      1.1  yamt 	STAILQ_ENTRY(tprof_buf) b_list;
     70      1.1  yamt 	tprof_sample_t b_data[];
     71      1.1  yamt } tprof_buf_t;
     72      1.1  yamt #define	TPROF_BUF_BYTESIZE(sz) \
     73      1.1  yamt 	(sizeof(tprof_buf_t) + (sz) * sizeof(tprof_sample_t))
     74      1.1  yamt #define	TPROF_MAX_SAMPLES_PER_BUF	10000
     75      1.1  yamt 
     76      1.1  yamt #define	TPROF_MAX_BUF			100
     77      1.1  yamt 
     78      1.1  yamt typedef struct {
     79      1.1  yamt 	tprof_buf_t *c_buf;
     80      1.1  yamt 	struct work c_work;
     81      1.1  yamt 	callout_t c_callout;
     82      1.1  yamt } __aligned(CACHE_LINE_SIZE) tprof_cpu_t;
     83      1.1  yamt 
     84  1.3.2.1   jym typedef struct tprof_backend {
     85  1.3.2.1   jym 	const char *tb_name;
     86  1.3.2.1   jym 	const tprof_backend_ops_t *tb_ops;
     87  1.3.2.1   jym 	LIST_ENTRY(tprof_backend) tb_list;
     88  1.3.2.1   jym 	int tb_usecount;	/* S: */
     89  1.3.2.1   jym } tprof_backend_t;
     90      1.3  yamt 
     91      1.1  yamt static kmutex_t tprof_lock;
     92      1.1  yamt static bool tprof_running;
     93  1.3.2.1   jym static u_int tprof_nworker;		/* L: # of running worker LWPs */
     94      1.1  yamt static lwp_t *tprof_owner;
     95  1.3.2.1   jym static STAILQ_HEAD(, tprof_buf) tprof_list; /* L: global buffer list */
     96  1.3.2.1   jym static u_int tprof_nbuf_on_list;	/* L: # of buffers on tprof_list */
     97      1.1  yamt static struct workqueue *tprof_wq;
     98      1.1  yamt static tprof_cpu_t tprof_cpus[MAXCPUS] __aligned(CACHE_LINE_SIZE);
     99      1.1  yamt static u_int tprof_samples_per_buf;
    100      1.1  yamt 
    101  1.3.2.1   jym static tprof_backend_t *tprof_backend;	/* S: */
    102  1.3.2.1   jym static LIST_HEAD(, tprof_backend) tprof_backends =
    103  1.3.2.1   jym     LIST_HEAD_INITIALIZER(tprof_backend); /* S: */
    104  1.3.2.1   jym 
    105      1.1  yamt static kmutex_t tprof_reader_lock;
    106  1.3.2.1   jym static kcondvar_t tprof_reader_cv;	/* L: */
    107  1.3.2.1   jym static off_t tprof_reader_offset;	/* R: */
    108      1.1  yamt 
    109      1.1  yamt static kmutex_t tprof_startstop_lock;
    110  1.3.2.1   jym static kcondvar_t tprof_cv;		/* L: */
    111      1.1  yamt 
    112  1.3.2.1   jym static struct tprof_stat tprof_stat;	/* L: */
    113      1.1  yamt 
    114      1.1  yamt static tprof_cpu_t *
    115      1.1  yamt tprof_cpu(struct cpu_info *ci)
    116      1.1  yamt {
    117      1.1  yamt 
    118      1.1  yamt 	return &tprof_cpus[cpu_index(ci)];
    119      1.1  yamt }
    120      1.1  yamt 
    121      1.1  yamt static tprof_cpu_t *
    122      1.1  yamt tprof_curcpu(void)
    123      1.1  yamt {
    124      1.1  yamt 
    125      1.1  yamt 	return tprof_cpu(curcpu());
    126      1.1  yamt }
    127      1.1  yamt 
    128      1.1  yamt static tprof_buf_t *
    129      1.1  yamt tprof_buf_alloc(void)
    130      1.1  yamt {
    131      1.1  yamt 	tprof_buf_t *new;
    132      1.1  yamt 	u_int size = tprof_samples_per_buf;
    133      1.1  yamt 
    134      1.1  yamt 	new = kmem_alloc(TPROF_BUF_BYTESIZE(size), KM_SLEEP);
    135      1.1  yamt 	new->b_used = 0;
    136      1.1  yamt 	new->b_size = size;
    137      1.1  yamt 	new->b_overflow = 0;
    138      1.1  yamt 	return new;
    139      1.1  yamt }
    140      1.1  yamt 
    141      1.1  yamt static void
    142      1.1  yamt tprof_buf_free(tprof_buf_t *buf)
    143      1.1  yamt {
    144      1.1  yamt 
    145      1.1  yamt 	kmem_free(buf, TPROF_BUF_BYTESIZE(buf->b_size));
    146      1.1  yamt }
    147      1.1  yamt 
    148      1.1  yamt static tprof_buf_t *
    149      1.1  yamt tprof_buf_switch(tprof_cpu_t *c, tprof_buf_t *new)
    150      1.1  yamt {
    151      1.1  yamt 	tprof_buf_t *old;
    152      1.1  yamt 
    153      1.1  yamt 	old = c->c_buf;
    154      1.1  yamt 	c->c_buf = new;
    155      1.1  yamt 	return old;
    156      1.1  yamt }
    157      1.1  yamt 
    158      1.1  yamt static tprof_buf_t *
    159      1.1  yamt tprof_buf_refresh(void)
    160      1.1  yamt {
    161      1.1  yamt 	tprof_cpu_t * const c = tprof_curcpu();
    162      1.1  yamt 	tprof_buf_t *new;
    163      1.1  yamt 
    164      1.1  yamt 	new = tprof_buf_alloc();
    165      1.1  yamt 	return tprof_buf_switch(c, new);
    166      1.1  yamt }
    167      1.1  yamt 
    168      1.1  yamt static void
    169      1.1  yamt tprof_worker(struct work *wk, void *dummy)
    170      1.1  yamt {
    171      1.1  yamt 	tprof_cpu_t * const c = tprof_curcpu();
    172      1.1  yamt 	tprof_buf_t *buf;
    173      1.1  yamt 	bool shouldstop;
    174      1.1  yamt 
    175      1.1  yamt 	KASSERT(wk == &c->c_work);
    176      1.1  yamt 	KASSERT(dummy == NULL);
    177      1.1  yamt 
    178      1.1  yamt 	/*
    179      1.1  yamt 	 * get a per cpu buffer.
    180      1.1  yamt 	 */
    181      1.1  yamt 	buf = tprof_buf_refresh();
    182      1.1  yamt 
    183      1.1  yamt 	/*
    184      1.1  yamt 	 * and put it on the global list for read(2).
    185      1.1  yamt 	 */
    186      1.1  yamt 	mutex_enter(&tprof_lock);
    187      1.1  yamt 	shouldstop = !tprof_running;
    188      1.1  yamt 	if (shouldstop) {
    189      1.1  yamt 		KASSERT(tprof_nworker > 0);
    190      1.1  yamt 		tprof_nworker--;
    191      1.1  yamt 		cv_broadcast(&tprof_cv);
    192      1.1  yamt 		cv_broadcast(&tprof_reader_cv);
    193      1.1  yamt 	}
    194      1.1  yamt 	if (buf->b_used == 0) {
    195      1.1  yamt 		tprof_stat.ts_emptybuf++;
    196      1.1  yamt 	} else if (tprof_nbuf_on_list < TPROF_MAX_BUF) {
    197      1.1  yamt 		tprof_stat.ts_sample += buf->b_used;
    198      1.1  yamt 		tprof_stat.ts_overflow += buf->b_overflow;
    199      1.1  yamt 		tprof_stat.ts_buf++;
    200      1.1  yamt 		STAILQ_INSERT_TAIL(&tprof_list, buf, b_list);
    201      1.1  yamt 		tprof_nbuf_on_list++;
    202      1.1  yamt 		buf = NULL;
    203      1.1  yamt 		cv_broadcast(&tprof_reader_cv);
    204      1.1  yamt 	} else {
    205      1.1  yamt 		tprof_stat.ts_dropbuf_sample += buf->b_used;
    206      1.1  yamt 		tprof_stat.ts_dropbuf++;
    207      1.1  yamt 	}
    208      1.1  yamt 	mutex_exit(&tprof_lock);
    209      1.1  yamt 	if (buf) {
    210      1.1  yamt 		tprof_buf_free(buf);
    211      1.1  yamt 	}
    212      1.1  yamt 	if (!shouldstop) {
    213      1.1  yamt 		callout_schedule(&c->c_callout, hz);
    214      1.1  yamt 	}
    215      1.1  yamt }
    216      1.1  yamt 
    217      1.1  yamt static void
    218      1.1  yamt tprof_kick(void *vp)
    219      1.1  yamt {
    220      1.1  yamt 	struct cpu_info * const ci = vp;
    221      1.1  yamt 	tprof_cpu_t * const c = tprof_cpu(ci);
    222      1.1  yamt 
    223      1.1  yamt 	workqueue_enqueue(tprof_wq, &c->c_work, ci);
    224      1.1  yamt }
    225      1.1  yamt 
    226      1.1  yamt static void
    227      1.1  yamt tprof_stop1(void)
    228      1.1  yamt {
    229      1.1  yamt 	CPU_INFO_ITERATOR cii;
    230      1.1  yamt 	struct cpu_info *ci;
    231      1.1  yamt 
    232      1.1  yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    233  1.3.2.1   jym 	KASSERT(tprof_nworker == 0);
    234      1.1  yamt 
    235      1.1  yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    236      1.1  yamt 		tprof_cpu_t * const c = tprof_cpu(ci);
    237      1.1  yamt 		tprof_buf_t *old;
    238      1.1  yamt 
    239      1.1  yamt 		old = tprof_buf_switch(c, NULL);
    240      1.1  yamt 		if (old != NULL) {
    241      1.1  yamt 			tprof_buf_free(old);
    242      1.1  yamt 		}
    243      1.1  yamt 		callout_destroy(&c->c_callout);
    244      1.1  yamt 	}
    245      1.1  yamt 	workqueue_destroy(tprof_wq);
    246      1.1  yamt }
    247      1.1  yamt 
    248      1.1  yamt static int
    249      1.1  yamt tprof_start(const struct tprof_param *param)
    250      1.1  yamt {
    251      1.1  yamt 	CPU_INFO_ITERATOR cii;
    252      1.1  yamt 	struct cpu_info *ci;
    253      1.1  yamt 	int error;
    254      1.1  yamt 	uint64_t freq;
    255  1.3.2.1   jym 	tprof_backend_t *tb;
    256      1.1  yamt 
    257      1.1  yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    258      1.1  yamt 	if (tprof_running) {
    259      1.1  yamt 		error = EBUSY;
    260      1.1  yamt 		goto done;
    261      1.1  yamt 	}
    262      1.1  yamt 
    263  1.3.2.1   jym 	tb = tprof_backend;
    264  1.3.2.1   jym 	if (tb == NULL) {
    265  1.3.2.1   jym 		error = ENOENT;
    266  1.3.2.1   jym 		goto done;
    267  1.3.2.1   jym 	}
    268  1.3.2.1   jym 	if (tb->tb_usecount > 0) {
    269  1.3.2.1   jym 		error = EBUSY;
    270  1.3.2.1   jym 		goto done;
    271  1.3.2.1   jym 	}
    272  1.3.2.1   jym 
    273  1.3.2.1   jym 	tb->tb_usecount++;
    274  1.3.2.1   jym 	freq = tb->tb_ops->tbo_estimate_freq();
    275      1.1  yamt 	tprof_samples_per_buf = MIN(freq * 2, TPROF_MAX_SAMPLES_PER_BUF);
    276      1.1  yamt 
    277      1.1  yamt 	error = workqueue_create(&tprof_wq, "tprofmv", tprof_worker, NULL,
    278      1.2  yamt 	    PRI_NONE, IPL_SOFTCLOCK, WQ_MPSAFE | WQ_PERCPU);
    279      1.1  yamt 	if (error != 0) {
    280      1.1  yamt 		goto done;
    281      1.1  yamt 	}
    282      1.1  yamt 
    283      1.1  yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    284      1.1  yamt 		tprof_cpu_t * const c = tprof_cpu(ci);
    285      1.1  yamt 		tprof_buf_t *new;
    286      1.1  yamt 		tprof_buf_t *old;
    287      1.1  yamt 
    288      1.1  yamt 		new = tprof_buf_alloc();
    289      1.1  yamt 		old = tprof_buf_switch(c, new);
    290      1.1  yamt 		if (old != NULL) {
    291      1.1  yamt 			tprof_buf_free(old);
    292      1.1  yamt 		}
    293      1.1  yamt 		callout_init(&c->c_callout, CALLOUT_MPSAFE);
    294      1.1  yamt 		callout_setfunc(&c->c_callout, tprof_kick, ci);
    295      1.1  yamt 	}
    296      1.1  yamt 
    297  1.3.2.1   jym 	error = tb->tb_ops->tbo_start(NULL);
    298      1.1  yamt 	if (error != 0) {
    299      1.1  yamt 		tprof_stop1();
    300      1.1  yamt 		goto done;
    301      1.1  yamt 	}
    302      1.1  yamt 
    303      1.1  yamt 	mutex_enter(&tprof_lock);
    304      1.1  yamt 	tprof_running = true;
    305      1.1  yamt 	mutex_exit(&tprof_lock);
    306      1.1  yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    307      1.1  yamt 		tprof_cpu_t * const c = tprof_cpu(ci);
    308      1.1  yamt 
    309      1.1  yamt 		mutex_enter(&tprof_lock);
    310      1.1  yamt 		tprof_nworker++;
    311      1.1  yamt 		mutex_exit(&tprof_lock);
    312      1.1  yamt 		workqueue_enqueue(tprof_wq, &c->c_work, ci);
    313      1.1  yamt 	}
    314      1.1  yamt done:
    315      1.1  yamt 	return error;
    316      1.1  yamt }
    317      1.1  yamt 
    318      1.1  yamt static void
    319      1.1  yamt tprof_stop(void)
    320      1.1  yamt {
    321      1.1  yamt 	CPU_INFO_ITERATOR cii;
    322      1.1  yamt 	struct cpu_info *ci;
    323  1.3.2.1   jym 	tprof_backend_t *tb;
    324      1.1  yamt 
    325      1.1  yamt 	KASSERT(mutex_owned(&tprof_startstop_lock));
    326      1.1  yamt 	if (!tprof_running) {
    327      1.1  yamt 		goto done;
    328      1.1  yamt 	}
    329      1.1  yamt 
    330  1.3.2.1   jym 	tb = tprof_backend;
    331  1.3.2.1   jym 	KASSERT(tb->tb_usecount > 0);
    332  1.3.2.1   jym 	tb->tb_ops->tbo_stop(NULL);
    333  1.3.2.1   jym 	tb->tb_usecount--;
    334      1.1  yamt 
    335      1.1  yamt 	mutex_enter(&tprof_lock);
    336      1.1  yamt 	tprof_running = false;
    337      1.1  yamt 	cv_broadcast(&tprof_reader_cv);
    338      1.1  yamt 	mutex_exit(&tprof_lock);
    339      1.1  yamt 
    340      1.1  yamt 	for (CPU_INFO_FOREACH(cii, ci)) {
    341      1.1  yamt 		mutex_enter(&tprof_lock);
    342      1.1  yamt 		while (tprof_nworker > 0) {
    343      1.1  yamt 			cv_wait(&tprof_cv, &tprof_lock);
    344      1.1  yamt 		}
    345      1.1  yamt 		mutex_exit(&tprof_lock);
    346      1.1  yamt 	}
    347      1.1  yamt 
    348      1.1  yamt 	tprof_stop1();
    349      1.1  yamt done:
    350      1.1  yamt 	;
    351      1.1  yamt }
    352      1.1  yamt 
    353  1.3.2.1   jym /*
    354  1.3.2.1   jym  * tprof_clear: drain unread samples.
    355  1.3.2.1   jym  */
    356  1.3.2.1   jym 
    357      1.1  yamt static void
    358      1.1  yamt tprof_clear(void)
    359      1.1  yamt {
    360      1.1  yamt 	tprof_buf_t *buf;
    361      1.1  yamt 
    362      1.1  yamt 	mutex_enter(&tprof_reader_lock);
    363      1.1  yamt 	mutex_enter(&tprof_lock);
    364      1.1  yamt 	while ((buf = STAILQ_FIRST(&tprof_list)) != NULL) {
    365      1.1  yamt 		if (buf != NULL) {
    366      1.1  yamt 			STAILQ_REMOVE_HEAD(&tprof_list, b_list);
    367      1.1  yamt 			KASSERT(tprof_nbuf_on_list > 0);
    368      1.1  yamt 			tprof_nbuf_on_list--;
    369      1.1  yamt 			mutex_exit(&tprof_lock);
    370      1.1  yamt 			tprof_buf_free(buf);
    371      1.1  yamt 			mutex_enter(&tprof_lock);
    372      1.1  yamt 		}
    373      1.1  yamt 	}
    374      1.1  yamt 	KASSERT(tprof_nbuf_on_list == 0);
    375      1.1  yamt 	mutex_exit(&tprof_lock);
    376      1.1  yamt 	tprof_reader_offset = 0;
    377      1.1  yamt 	mutex_exit(&tprof_reader_lock);
    378      1.1  yamt 
    379      1.1  yamt 	memset(&tprof_stat, 0, sizeof(tprof_stat));
    380      1.1  yamt }
    381      1.1  yamt 
    382  1.3.2.1   jym static tprof_backend_t *
    383  1.3.2.1   jym tprof_backend_lookup(const char *name)
    384  1.3.2.1   jym {
    385  1.3.2.1   jym 	tprof_backend_t *tb;
    386  1.3.2.1   jym 
    387  1.3.2.1   jym 	KASSERT(mutex_owned(&tprof_startstop_lock));
    388  1.3.2.1   jym 
    389  1.3.2.1   jym 	LIST_FOREACH(tb, &tprof_backends, tb_list) {
    390  1.3.2.1   jym 		if (!strcmp(tb->tb_name, name)) {
    391  1.3.2.1   jym 			return tb;
    392  1.3.2.1   jym 		}
    393  1.3.2.1   jym 	}
    394  1.3.2.1   jym 	return NULL;
    395  1.3.2.1   jym }
    396  1.3.2.1   jym 
    397      1.1  yamt /* -------------------- backend interfaces */
    398      1.1  yamt 
    399      1.1  yamt /*
    400      1.1  yamt  * tprof_sample: record a sample on the per-cpu buffer.
    401      1.1  yamt  *
    402      1.1  yamt  * be careful; can be called in NMI context.
    403      1.1  yamt  * we are assuming that curcpu() is safe.
    404      1.1  yamt  */
    405      1.1  yamt 
    406      1.1  yamt void
    407  1.3.2.1   jym tprof_sample(tprof_backend_cookie_t *cookie, const tprof_frame_info_t *tfi)
    408      1.1  yamt {
    409      1.1  yamt 	tprof_cpu_t * const c = tprof_curcpu();
    410      1.1  yamt 	tprof_buf_t * const buf = c->c_buf;
    411  1.3.2.1   jym 	const uintptr_t pc = tfi->tfi_pc;
    412      1.1  yamt 	u_int idx;
    413      1.1  yamt 
    414      1.1  yamt 	idx = buf->b_used;
    415      1.1  yamt 	if (__predict_false(idx >= buf->b_size)) {
    416      1.1  yamt 		buf->b_overflow++;
    417      1.1  yamt 		return;
    418      1.1  yamt 	}
    419      1.1  yamt 	buf->b_data[idx].s_pc = pc;
    420      1.1  yamt 	buf->b_used = idx + 1;
    421      1.1  yamt }
    422      1.1  yamt 
    423  1.3.2.1   jym /*
    424  1.3.2.1   jym  * tprof_backend_register:
    425  1.3.2.1   jym  */
    426  1.3.2.1   jym 
    427  1.3.2.1   jym int
    428  1.3.2.1   jym tprof_backend_register(const char *name, const tprof_backend_ops_t *ops,
    429  1.3.2.1   jym     int vers)
    430  1.3.2.1   jym {
    431  1.3.2.1   jym 	tprof_backend_t *tb;
    432  1.3.2.1   jym 
    433  1.3.2.1   jym 	if (vers != TPROF_BACKEND_VERSION) {
    434  1.3.2.1   jym 		return EINVAL;
    435  1.3.2.1   jym 	}
    436  1.3.2.1   jym 
    437  1.3.2.1   jym 	mutex_enter(&tprof_startstop_lock);
    438  1.3.2.1   jym 	tb = tprof_backend_lookup(name);
    439  1.3.2.1   jym 	if (tb != NULL) {
    440  1.3.2.1   jym 		mutex_exit(&tprof_startstop_lock);
    441  1.3.2.1   jym 		return EEXIST;
    442  1.3.2.1   jym 	}
    443  1.3.2.1   jym #if 1 /* XXX for now */
    444  1.3.2.1   jym 	if (!LIST_EMPTY(&tprof_backends)) {
    445  1.3.2.1   jym 		mutex_exit(&tprof_startstop_lock);
    446  1.3.2.1   jym 		return ENOTSUP;
    447  1.3.2.1   jym 	}
    448  1.3.2.1   jym #endif
    449  1.3.2.1   jym 	tb = kmem_alloc(sizeof(*tb), KM_SLEEP);
    450  1.3.2.1   jym 	tb->tb_name = name;
    451  1.3.2.1   jym 	tb->tb_ops = ops;
    452  1.3.2.1   jym 	tb->tb_usecount = 0;
    453  1.3.2.1   jym 	LIST_INSERT_HEAD(&tprof_backends, tb, tb_list);
    454  1.3.2.1   jym #if 1 /* XXX for now */
    455  1.3.2.1   jym 	if (tprof_backend == NULL) {
    456  1.3.2.1   jym 		tprof_backend = tb;
    457  1.3.2.1   jym 	}
    458  1.3.2.1   jym #endif
    459  1.3.2.1   jym 	mutex_exit(&tprof_startstop_lock);
    460  1.3.2.1   jym 
    461  1.3.2.1   jym 	return 0;
    462  1.3.2.1   jym }
    463  1.3.2.1   jym 
    464  1.3.2.1   jym /*
    465  1.3.2.1   jym  * tprof_backend_unregister:
    466  1.3.2.1   jym  */
    467  1.3.2.1   jym 
    468  1.3.2.1   jym int
    469  1.3.2.1   jym tprof_backend_unregister(const char *name)
    470  1.3.2.1   jym {
    471  1.3.2.1   jym 	tprof_backend_t *tb;
    472  1.3.2.1   jym 
    473  1.3.2.1   jym 	mutex_enter(&tprof_startstop_lock);
    474  1.3.2.1   jym 	tb = tprof_backend_lookup(name);
    475  1.3.2.1   jym #if defined(DIAGNOSTIC)
    476  1.3.2.1   jym 	if (tb == NULL) {
    477  1.3.2.1   jym 		mutex_exit(&tprof_startstop_lock);
    478  1.3.2.1   jym 		panic("%s: not found '%s'", __func__, name);
    479  1.3.2.1   jym 	}
    480  1.3.2.1   jym #endif /* defined(DIAGNOSTIC) */
    481  1.3.2.1   jym 	if (tb->tb_usecount > 0) {
    482  1.3.2.1   jym 		mutex_exit(&tprof_startstop_lock);
    483  1.3.2.1   jym 		return EBUSY;
    484  1.3.2.1   jym 	}
    485  1.3.2.1   jym #if 1 /* XXX for now */
    486  1.3.2.1   jym 	if (tprof_backend == tb) {
    487  1.3.2.1   jym 		tprof_backend = NULL;
    488  1.3.2.1   jym 	}
    489  1.3.2.1   jym #endif
    490  1.3.2.1   jym 	LIST_REMOVE(tb, tb_list);
    491  1.3.2.1   jym 	mutex_exit(&tprof_startstop_lock);
    492  1.3.2.1   jym 
    493  1.3.2.1   jym 	kmem_free(tb, sizeof(*tb));
    494  1.3.2.1   jym 
    495  1.3.2.1   jym 	return 0;
    496  1.3.2.1   jym }
    497  1.3.2.1   jym 
    498      1.1  yamt /* -------------------- cdevsw interfaces */
    499      1.1  yamt 
    500      1.1  yamt void tprofattach(int);
    501      1.1  yamt 
    502      1.1  yamt static int
    503      1.1  yamt tprof_open(dev_t dev, int flags, int type, struct lwp *l)
    504      1.1  yamt {
    505      1.1  yamt 
    506      1.1  yamt 	if (minor(dev) != 0) {
    507      1.1  yamt 		return EXDEV;
    508      1.1  yamt 	}
    509      1.1  yamt 	mutex_enter(&tprof_lock);
    510      1.1  yamt 	if (tprof_owner != NULL) {
    511      1.1  yamt 		mutex_exit(&tprof_lock);
    512      1.1  yamt 		return  EBUSY;
    513      1.1  yamt 	}
    514      1.1  yamt 	tprof_owner = curlwp;
    515      1.1  yamt 	mutex_exit(&tprof_lock);
    516      1.1  yamt 
    517      1.1  yamt 	return 0;
    518      1.1  yamt }
    519      1.1  yamt 
    520      1.1  yamt static int
    521      1.1  yamt tprof_close(dev_t dev, int flags, int type, struct lwp *l)
    522      1.1  yamt {
    523      1.1  yamt 
    524      1.1  yamt 	KASSERT(minor(dev) == 0);
    525      1.1  yamt 
    526      1.1  yamt 	mutex_enter(&tprof_startstop_lock);
    527      1.1  yamt 	mutex_enter(&tprof_lock);
    528      1.1  yamt 	tprof_owner = NULL;
    529      1.1  yamt 	mutex_exit(&tprof_lock);
    530      1.1  yamt 	tprof_stop();
    531      1.1  yamt 	tprof_clear();
    532      1.1  yamt 	mutex_exit(&tprof_startstop_lock);
    533      1.1  yamt 
    534      1.1  yamt 	return 0;
    535      1.1  yamt }
    536      1.1  yamt 
    537      1.1  yamt static int
    538      1.1  yamt tprof_read(dev_t dev, struct uio *uio, int flags)
    539      1.1  yamt {
    540      1.1  yamt 	tprof_buf_t *buf;
    541      1.1  yamt 	size_t bytes;
    542      1.1  yamt 	size_t resid;
    543      1.1  yamt 	size_t done;
    544      1.1  yamt 	int error = 0;
    545      1.1  yamt 
    546      1.1  yamt 	KASSERT(minor(dev) == 0);
    547      1.1  yamt 	mutex_enter(&tprof_reader_lock);
    548      1.1  yamt 	while (uio->uio_resid > 0 && error == 0) {
    549      1.1  yamt 		/*
    550      1.1  yamt 		 * take the first buffer from the list.
    551      1.1  yamt 		 */
    552      1.1  yamt 		mutex_enter(&tprof_lock);
    553      1.1  yamt 		buf = STAILQ_FIRST(&tprof_list);
    554      1.1  yamt 		if (buf == NULL) {
    555      1.1  yamt 			if (tprof_nworker == 0) {
    556      1.1  yamt 				mutex_exit(&tprof_lock);
    557      1.1  yamt 				error = 0;
    558      1.1  yamt 				break;
    559      1.1  yamt 			}
    560      1.1  yamt 			mutex_exit(&tprof_reader_lock);
    561      1.1  yamt 			error = cv_wait_sig(&tprof_reader_cv, &tprof_lock);
    562      1.1  yamt 			mutex_exit(&tprof_lock);
    563      1.1  yamt 			mutex_enter(&tprof_reader_lock);
    564      1.1  yamt 			continue;
    565      1.1  yamt 		}
    566      1.1  yamt 		STAILQ_REMOVE_HEAD(&tprof_list, b_list);
    567      1.1  yamt 		KASSERT(tprof_nbuf_on_list > 0);
    568      1.1  yamt 		tprof_nbuf_on_list--;
    569      1.1  yamt 		mutex_exit(&tprof_lock);
    570      1.1  yamt 
    571      1.1  yamt 		/*
    572      1.1  yamt 		 * copy it out.
    573      1.1  yamt 		 */
    574      1.1  yamt 		bytes = MIN(buf->b_used * sizeof(tprof_sample_t) -
    575      1.1  yamt 		    tprof_reader_offset, uio->uio_resid);
    576      1.1  yamt 		resid = uio->uio_resid;
    577      1.1  yamt 		error = uiomove((char *)buf->b_data + tprof_reader_offset,
    578      1.1  yamt 		    bytes, uio);
    579      1.1  yamt 		done = resid - uio->uio_resid;
    580      1.1  yamt 		tprof_reader_offset += done;
    581      1.1  yamt 
    582      1.1  yamt 		/*
    583      1.1  yamt 		 * if we didn't consume the whole buffer,
    584      1.1  yamt 		 * put it back to the list.
    585      1.1  yamt 		 */
    586      1.1  yamt 		if (tprof_reader_offset <
    587      1.1  yamt 		    buf->b_used * sizeof(tprof_sample_t)) {
    588      1.1  yamt 			mutex_enter(&tprof_lock);
    589      1.1  yamt 			STAILQ_INSERT_HEAD(&tprof_list, buf, b_list);
    590      1.1  yamt 			tprof_nbuf_on_list++;
    591      1.1  yamt 			cv_broadcast(&tprof_reader_cv);
    592      1.1  yamt 			mutex_exit(&tprof_lock);
    593      1.1  yamt 		} else {
    594      1.1  yamt 			tprof_buf_free(buf);
    595      1.1  yamt 			tprof_reader_offset = 0;
    596      1.1  yamt 		}
    597      1.1  yamt 	}
    598      1.1  yamt 	mutex_exit(&tprof_reader_lock);
    599      1.1  yamt 
    600      1.1  yamt 	return error;
    601      1.1  yamt }
    602      1.1  yamt 
    603      1.1  yamt static int
    604      1.1  yamt tprof_ioctl(dev_t dev, u_long cmd, void *data, int flags, struct lwp *l)
    605      1.1  yamt {
    606      1.1  yamt 	const struct tprof_param *param;
    607      1.1  yamt 	int error = 0;
    608      1.1  yamt 
    609      1.1  yamt 	KASSERT(minor(dev) == 0);
    610      1.1  yamt 
    611      1.1  yamt 	switch (cmd) {
    612      1.1  yamt 	case TPROF_IOC_GETVERSION:
    613      1.1  yamt 		*(int *)data = TPROF_VERSION;
    614      1.1  yamt 		break;
    615      1.1  yamt 	case TPROF_IOC_START:
    616      1.1  yamt 		param = data;
    617      1.1  yamt 		mutex_enter(&tprof_startstop_lock);
    618      1.1  yamt 		error = tprof_start(param);
    619      1.1  yamt 		mutex_exit(&tprof_startstop_lock);
    620      1.1  yamt 		break;
    621      1.1  yamt 	case TPROF_IOC_STOP:
    622      1.1  yamt 		mutex_enter(&tprof_startstop_lock);
    623      1.1  yamt 		tprof_stop();
    624      1.1  yamt 		mutex_exit(&tprof_startstop_lock);
    625      1.1  yamt 		break;
    626      1.1  yamt 	case TPROF_IOC_GETSTAT:
    627      1.1  yamt 		mutex_enter(&tprof_lock);
    628      1.1  yamt 		memcpy(data, &tprof_stat, sizeof(tprof_stat));
    629      1.1  yamt 		mutex_exit(&tprof_lock);
    630      1.1  yamt 		break;
    631      1.1  yamt 	default:
    632      1.1  yamt 		error = EINVAL;
    633      1.1  yamt 		break;
    634      1.1  yamt 	}
    635      1.1  yamt 
    636      1.1  yamt 	return error;
    637      1.1  yamt }
    638      1.1  yamt 
    639      1.1  yamt const struct cdevsw tprof_cdevsw = {
    640      1.1  yamt 	.d_open = tprof_open,
    641      1.1  yamt 	.d_close = tprof_close,
    642      1.1  yamt 	.d_read = tprof_read,
    643      1.1  yamt 	.d_write = nowrite,
    644      1.1  yamt 	.d_ioctl = tprof_ioctl,
    645      1.1  yamt 	.d_stop = nostop,
    646      1.1  yamt 	.d_tty = notty,
    647      1.1  yamt 	.d_poll = nopoll,
    648      1.1  yamt 	.d_mmap = nommap,
    649      1.1  yamt 	.d_kqfilter = nokqfilter,
    650      1.1  yamt 	.d_flag = D_OTHER | D_MPSAFE,
    651      1.1  yamt };
    652      1.1  yamt 
    653      1.1  yamt void
    654      1.1  yamt tprofattach(int nunits)
    655      1.1  yamt {
    656      1.1  yamt 
    657  1.3.2.1   jym 	/* nothing */
    658  1.3.2.1   jym }
    659  1.3.2.1   jym 
    660  1.3.2.1   jym MODULE(MODULE_CLASS_DRIVER, tprof, NULL);
    661  1.3.2.1   jym 
    662  1.3.2.1   jym static void
    663  1.3.2.1   jym tprof_driver_init(void)
    664  1.3.2.1   jym {
    665  1.3.2.1   jym 
    666      1.1  yamt 	mutex_init(&tprof_lock, MUTEX_DEFAULT, IPL_NONE);
    667      1.1  yamt 	mutex_init(&tprof_reader_lock, MUTEX_DEFAULT, IPL_NONE);
    668      1.1  yamt 	mutex_init(&tprof_startstop_lock, MUTEX_DEFAULT, IPL_NONE);
    669      1.1  yamt 	cv_init(&tprof_cv, "tprof");
    670      1.1  yamt 	cv_init(&tprof_reader_cv, "tprofread");
    671      1.1  yamt 	STAILQ_INIT(&tprof_list);
    672      1.1  yamt }
    673  1.3.2.1   jym 
    674  1.3.2.1   jym static void
    675  1.3.2.1   jym tprof_driver_fini(void)
    676  1.3.2.1   jym {
    677  1.3.2.1   jym 
    678  1.3.2.1   jym 	mutex_destroy(&tprof_lock);
    679  1.3.2.1   jym 	mutex_destroy(&tprof_reader_lock);
    680  1.3.2.1   jym 	mutex_destroy(&tprof_startstop_lock);
    681  1.3.2.1   jym 	cv_destroy(&tprof_cv);
    682  1.3.2.1   jym 	cv_destroy(&tprof_reader_cv);
    683  1.3.2.1   jym }
    684  1.3.2.1   jym 
    685  1.3.2.1   jym static int
    686  1.3.2.1   jym tprof_modcmd(modcmd_t cmd, void *arg)
    687  1.3.2.1   jym {
    688  1.3.2.1   jym 
    689  1.3.2.1   jym 	switch (cmd) {
    690  1.3.2.1   jym 	case MODULE_CMD_INIT:
    691  1.3.2.1   jym 		tprof_driver_init();
    692  1.3.2.1   jym #if defined(_MODULE)
    693  1.3.2.1   jym 		{
    694  1.3.2.1   jym 			devmajor_t bmajor = NODEVMAJOR;
    695  1.3.2.1   jym 			devmajor_t cmajor = NODEVMAJOR;
    696  1.3.2.1   jym 			int error;
    697  1.3.2.1   jym 
    698  1.3.2.1   jym 			error = devsw_attach("tprof", NULL, &bmajor,
    699  1.3.2.1   jym 			    &tprof_cdevsw, &cmajor);
    700  1.3.2.1   jym 			if (error) {
    701  1.3.2.1   jym 				tprof_driver_fini();
    702  1.3.2.1   jym 				return error;
    703  1.3.2.1   jym 			}
    704  1.3.2.1   jym 		}
    705  1.3.2.1   jym #endif /* defined(_MODULE) */
    706  1.3.2.1   jym 		return 0;
    707  1.3.2.1   jym 
    708  1.3.2.1   jym 	case MODULE_CMD_FINI:
    709  1.3.2.1   jym #if defined(_MODULE)
    710  1.3.2.1   jym 		{
    711  1.3.2.1   jym 			int error;
    712  1.3.2.1   jym 			error = devsw_detach(NULL, &tprof_cdevsw);
    713  1.3.2.1   jym 			if (error) {
    714  1.3.2.1   jym 				return error;
    715  1.3.2.1   jym 			}
    716  1.3.2.1   jym 		}
    717  1.3.2.1   jym #endif /* defined(_MODULE) */
    718  1.3.2.1   jym 		tprof_driver_fini();
    719  1.3.2.1   jym 		return 0;
    720  1.3.2.1   jym 
    721  1.3.2.1   jym 	default:
    722  1.3.2.1   jym 		return ENOTTY;
    723  1.3.2.1   jym 	}
    724  1.3.2.1   jym }
    725