subr_workqueue.c revision 1.3.6.4 1 1.3.6.4 yamt /* $NetBSD: subr_workqueue.c,v 1.3.6.4 2007/02/26 09:11:16 yamt Exp $ */
2 1.3.6.2 yamt
3 1.3.6.2 yamt /*-
4 1.3.6.2 yamt * Copyright (c)2002, 2005 YAMAMOTO Takashi,
5 1.3.6.2 yamt * All rights reserved.
6 1.3.6.2 yamt *
7 1.3.6.2 yamt * Redistribution and use in source and binary forms, with or without
8 1.3.6.2 yamt * modification, are permitted provided that the following conditions
9 1.3.6.2 yamt * are met:
10 1.3.6.2 yamt * 1. Redistributions of source code must retain the above copyright
11 1.3.6.2 yamt * notice, this list of conditions and the following disclaimer.
12 1.3.6.2 yamt * 2. Redistributions in binary form must reproduce the above copyright
13 1.3.6.2 yamt * notice, this list of conditions and the following disclaimer in the
14 1.3.6.2 yamt * documentation and/or other materials provided with the distribution.
15 1.3.6.2 yamt *
16 1.3.6.2 yamt * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17 1.3.6.2 yamt * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18 1.3.6.2 yamt * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19 1.3.6.2 yamt * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20 1.3.6.2 yamt * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21 1.3.6.2 yamt * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22 1.3.6.2 yamt * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23 1.3.6.2 yamt * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24 1.3.6.2 yamt * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25 1.3.6.2 yamt * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26 1.3.6.2 yamt * SUCH DAMAGE.
27 1.3.6.2 yamt */
28 1.3.6.2 yamt
29 1.3.6.2 yamt #include <sys/cdefs.h>
30 1.3.6.4 yamt __KERNEL_RCSID(0, "$NetBSD: subr_workqueue.c,v 1.3.6.4 2007/02/26 09:11:16 yamt Exp $");
31 1.3.6.2 yamt
32 1.3.6.2 yamt #include <sys/param.h>
33 1.3.6.2 yamt #include <sys/systm.h>
34 1.3.6.2 yamt #include <sys/kthread.h>
35 1.3.6.3 yamt #include <sys/kmem.h>
36 1.3.6.2 yamt #include <sys/proc.h>
37 1.3.6.2 yamt #include <sys/workqueue.h>
38 1.3.6.4 yamt #include <sys/mutex.h>
39 1.3.6.4 yamt #include <sys/condvar.h>
40 1.3.6.2 yamt
41 1.3.6.2 yamt SIMPLEQ_HEAD(workqhead, work);
42 1.3.6.2 yamt
43 1.3.6.2 yamt struct workqueue_queue {
44 1.3.6.4 yamt kmutex_t q_mutex;
45 1.3.6.4 yamt kcondvar_t q_cv;
46 1.3.6.2 yamt struct workqhead q_queue;
47 1.3.6.2 yamt struct proc *q_worker;
48 1.3.6.2 yamt };
49 1.3.6.2 yamt
50 1.3.6.2 yamt struct workqueue {
51 1.3.6.2 yamt struct workqueue_queue wq_queue; /* todo: make this per-cpu */
52 1.3.6.2 yamt
53 1.3.6.2 yamt void (*wq_func)(struct work *, void *);
54 1.3.6.2 yamt void *wq_arg;
55 1.3.6.2 yamt const char *wq_name;
56 1.3.6.2 yamt int wq_prio;
57 1.3.6.3 yamt ipl_cookie_t wq_ipl;
58 1.3.6.2 yamt };
59 1.3.6.2 yamt
60 1.3.6.2 yamt #define POISON 0xaabbccdd
61 1.3.6.2 yamt
62 1.3.6.2 yamt static void
63 1.3.6.2 yamt workqueue_runlist(struct workqueue *wq, struct workqhead *list)
64 1.3.6.2 yamt {
65 1.3.6.2 yamt struct work *wk;
66 1.3.6.2 yamt struct work *next;
67 1.3.6.2 yamt
68 1.3.6.2 yamt /*
69 1.3.6.2 yamt * note that "list" is not a complete SIMPLEQ.
70 1.3.6.2 yamt */
71 1.3.6.2 yamt
72 1.3.6.2 yamt for (wk = SIMPLEQ_FIRST(list); wk != NULL; wk = next) {
73 1.3.6.2 yamt next = SIMPLEQ_NEXT(wk, wk_entry);
74 1.3.6.2 yamt (*wq->wq_func)(wk, wq->wq_arg);
75 1.3.6.2 yamt }
76 1.3.6.2 yamt }
77 1.3.6.2 yamt
78 1.3.6.2 yamt static void
79 1.3.6.2 yamt workqueue_run(struct workqueue *wq)
80 1.3.6.2 yamt {
81 1.3.6.2 yamt struct workqueue_queue *q = &wq->wq_queue;
82 1.3.6.2 yamt
83 1.3.6.2 yamt for (;;) {
84 1.3.6.2 yamt struct workqhead tmp;
85 1.3.6.2 yamt
86 1.3.6.2 yamt /*
87 1.3.6.2 yamt * we violate abstraction of SIMPLEQ.
88 1.3.6.2 yamt */
89 1.3.6.2 yamt
90 1.3.6.2 yamt #if defined(DIAGNOSTIC)
91 1.3.6.2 yamt tmp.sqh_last = (void *)POISON;
92 1.3.6.2 yamt #endif /* defined(DIAGNOSTIC) */
93 1.3.6.2 yamt
94 1.3.6.4 yamt mutex_enter(&q->q_mutex);
95 1.3.6.4 yamt while (SIMPLEQ_EMPTY(&q->q_queue))
96 1.3.6.4 yamt cv_wait(&q->q_cv, &q->q_mutex);
97 1.3.6.2 yamt tmp.sqh_first = q->q_queue.sqh_first; /* XXX */
98 1.3.6.2 yamt SIMPLEQ_INIT(&q->q_queue);
99 1.3.6.4 yamt mutex_exit(&q->q_mutex);
100 1.3.6.2 yamt
101 1.3.6.2 yamt workqueue_runlist(wq, &tmp);
102 1.3.6.2 yamt }
103 1.3.6.2 yamt }
104 1.3.6.2 yamt
105 1.3.6.2 yamt static void
106 1.3.6.2 yamt workqueue_worker(void *arg)
107 1.3.6.2 yamt {
108 1.3.6.2 yamt struct workqueue *wq = arg;
109 1.3.6.4 yamt struct lwp *l;
110 1.3.6.4 yamt
111 1.3.6.4 yamt l = curlwp;
112 1.3.6.4 yamt lwp_lock(l);
113 1.3.6.4 yamt l->l_priority = wq->wq_prio;
114 1.3.6.4 yamt l->l_usrpri = wq->wq_prio;
115 1.3.6.4 yamt lwp_unlock(l);
116 1.3.6.2 yamt
117 1.3.6.2 yamt workqueue_run(wq);
118 1.3.6.2 yamt }
119 1.3.6.2 yamt
120 1.3.6.2 yamt static void
121 1.3.6.2 yamt workqueue_init(struct workqueue *wq, const char *name,
122 1.3.6.2 yamt void (*callback_func)(struct work *, void *), void *callback_arg,
123 1.3.6.2 yamt int prio, int ipl)
124 1.3.6.2 yamt {
125 1.3.6.2 yamt
126 1.3.6.3 yamt wq->wq_ipl = makeiplcookie(ipl);
127 1.3.6.2 yamt wq->wq_prio = prio;
128 1.3.6.2 yamt wq->wq_name = name;
129 1.3.6.2 yamt wq->wq_func = callback_func;
130 1.3.6.2 yamt wq->wq_arg = callback_arg;
131 1.3.6.2 yamt }
132 1.3.6.2 yamt
133 1.3.6.2 yamt static int
134 1.3.6.4 yamt workqueue_initqueue(struct workqueue *wq, int ipl)
135 1.3.6.2 yamt {
136 1.3.6.2 yamt struct workqueue_queue *q = &wq->wq_queue;
137 1.3.6.2 yamt int error;
138 1.3.6.2 yamt
139 1.3.6.4 yamt mutex_init(&q->q_mutex, MUTEX_SPIN, ipl);
140 1.3.6.4 yamt cv_init(&q->q_cv, wq->wq_name);
141 1.3.6.2 yamt SIMPLEQ_INIT(&q->q_queue);
142 1.3.6.2 yamt error = kthread_create1(workqueue_worker, wq, &q->q_worker,
143 1.3.6.2 yamt wq->wq_name);
144 1.3.6.2 yamt
145 1.3.6.2 yamt return error;
146 1.3.6.2 yamt }
147 1.3.6.2 yamt
148 1.3.6.3 yamt struct workqueue_exitargs {
149 1.3.6.3 yamt struct work wqe_wk;
150 1.3.6.3 yamt struct workqueue_queue *wqe_q;
151 1.3.6.3 yamt };
152 1.3.6.3 yamt
153 1.3.6.3 yamt static void
154 1.3.6.3 yamt workqueue_exit(struct work *wk, void *arg)
155 1.3.6.3 yamt {
156 1.3.6.3 yamt struct workqueue_exitargs *wqe = (void *)wk;
157 1.3.6.3 yamt struct workqueue_queue *q = wqe->wqe_q;
158 1.3.6.3 yamt
159 1.3.6.3 yamt /*
160 1.3.6.4 yamt * only competition at this point is workqueue_finiqueue.
161 1.3.6.3 yamt */
162 1.3.6.3 yamt
163 1.3.6.3 yamt KASSERT(q->q_worker == curproc);
164 1.3.6.4 yamt mutex_enter(&q->q_mutex);
165 1.3.6.3 yamt q->q_worker = NULL;
166 1.3.6.4 yamt cv_signal(&q->q_cv);
167 1.3.6.4 yamt mutex_exit(&q->q_mutex);
168 1.3.6.3 yamt kthread_exit(0);
169 1.3.6.3 yamt }
170 1.3.6.3 yamt
171 1.3.6.3 yamt static void
172 1.3.6.3 yamt workqueue_finiqueue(struct workqueue *wq)
173 1.3.6.3 yamt {
174 1.3.6.3 yamt struct workqueue_queue *q = &wq->wq_queue;
175 1.3.6.3 yamt struct workqueue_exitargs wqe;
176 1.3.6.3 yamt
177 1.3.6.3 yamt wq->wq_func = workqueue_exit;
178 1.3.6.3 yamt
179 1.3.6.3 yamt wqe.wqe_q = q;
180 1.3.6.3 yamt KASSERT(SIMPLEQ_EMPTY(&q->q_queue));
181 1.3.6.3 yamt KASSERT(q->q_worker != NULL);
182 1.3.6.4 yamt mutex_enter(&q->q_mutex);
183 1.3.6.3 yamt SIMPLEQ_INSERT_TAIL(&q->q_queue, &wqe.wqe_wk, wk_entry);
184 1.3.6.4 yamt cv_signal(&q->q_cv);
185 1.3.6.3 yamt while (q->q_worker != NULL) {
186 1.3.6.4 yamt cv_wait(&q->q_cv, &q->q_mutex);
187 1.3.6.3 yamt }
188 1.3.6.4 yamt mutex_exit(&q->q_mutex);
189 1.3.6.4 yamt mutex_destroy(&q->q_mutex);
190 1.3.6.4 yamt cv_destroy(&q->q_cv);
191 1.3.6.3 yamt }
192 1.3.6.3 yamt
193 1.3.6.2 yamt /* --- */
194 1.3.6.2 yamt
195 1.3.6.2 yamt int
196 1.3.6.2 yamt workqueue_create(struct workqueue **wqp, const char *name,
197 1.3.6.2 yamt void (*callback_func)(struct work *, void *), void *callback_arg,
198 1.3.6.2 yamt int prio, int ipl, int flags)
199 1.3.6.2 yamt {
200 1.3.6.2 yamt struct workqueue *wq;
201 1.3.6.2 yamt int error;
202 1.3.6.2 yamt
203 1.3.6.3 yamt wq = kmem_alloc(sizeof(*wq), KM_SLEEP);
204 1.3.6.2 yamt if (wq == NULL) {
205 1.3.6.2 yamt return ENOMEM;
206 1.3.6.2 yamt }
207 1.3.6.2 yamt
208 1.3.6.2 yamt workqueue_init(wq, name, callback_func, callback_arg, prio, ipl);
209 1.3.6.2 yamt
210 1.3.6.4 yamt error = workqueue_initqueue(wq, ipl);
211 1.3.6.2 yamt if (error) {
212 1.3.6.3 yamt kmem_free(wq, sizeof(*wq));
213 1.3.6.2 yamt return error;
214 1.3.6.2 yamt }
215 1.3.6.2 yamt
216 1.3.6.2 yamt *wqp = wq;
217 1.3.6.2 yamt return 0;
218 1.3.6.2 yamt }
219 1.3.6.2 yamt
220 1.3.6.2 yamt void
221 1.3.6.3 yamt workqueue_destroy(struct workqueue *wq)
222 1.3.6.3 yamt {
223 1.3.6.3 yamt
224 1.3.6.3 yamt workqueue_finiqueue(wq);
225 1.3.6.3 yamt kmem_free(wq, sizeof(*wq));
226 1.3.6.3 yamt }
227 1.3.6.3 yamt
228 1.3.6.3 yamt void
229 1.3.6.2 yamt workqueue_enqueue(struct workqueue *wq, struct work *wk)
230 1.3.6.2 yamt {
231 1.3.6.2 yamt struct workqueue_queue *q = &wq->wq_queue;
232 1.3.6.2 yamt
233 1.3.6.4 yamt mutex_enter(&q->q_mutex);
234 1.3.6.4 yamt if (SIMPLEQ_EMPTY(&q->q_queue))
235 1.3.6.4 yamt cv_signal(&q->q_cv);
236 1.3.6.2 yamt SIMPLEQ_INSERT_TAIL(&q->q_queue, wk, wk_entry);
237 1.3.6.4 yamt mutex_exit(&q->q_mutex);
238 1.3.6.2 yamt }
239