chfs_gc.c revision 1.2 1 /* $NetBSD: chfs_gc.c,v 1.2 2011/11/24 21:09:37 agc Exp $ */
2
3 /*-
4 * Copyright (c) 2010 Department of Software Engineering,
5 * University of Szeged, Hungary
6 * Copyright (c) 2010 Tamas Toth <ttoth (at) inf.u-szeged.hu>
7 * Copyright (c) 2010 Adam Hoka <ahoka (at) NetBSD.org>
8 * All rights reserved.
9 *
10 * This code is derived from software contributed to The NetBSD Foundation
11 * by the Department of Software Engineering, University of Szeged, Hungary
12 *
13 * Redistribution and use in source and binary forms, with or without
14 * modification, are permitted provided that the following conditions
15 * are met:
16 * 1. Redistributions of source code must retain the above copyright
17 * notice, this list of conditions and the following disclaimer.
18 * 2. Redistributions in binary form must reproduce the above copyright
19 * notice, this list of conditions and the following disclaimer in the
20 * documentation and/or other materials provided with the distribution.
21 *
22 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
23 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
24 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
25 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
26 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
27 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
28 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
29 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
30 * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
32 * SUCH DAMAGE.
33 */
34
35 #include "chfs.h"
36
37 void chfs_gc_release_inode(struct chfs_mount *,
38 struct chfs_inode *);
39 struct chfs_inode *chfs_gc_fetch_inode(struct chfs_mount *,
40 ino_t, uint32_t);
41 int chfs_check(struct chfs_mount *, struct chfs_vnode_cache *);
42 void chfs_clear_inode(struct chfs_mount *, struct chfs_inode *);
43
44
45 struct chfs_eraseblock *find_gc_block(struct chfs_mount *);
46 int chfs_gcollect_pristine(struct chfs_mount *,
47 struct chfs_eraseblock *,
48 struct chfs_vnode_cache *, struct chfs_node_ref *);
49 int chfs_gcollect_live(struct chfs_mount *,
50 struct chfs_eraseblock *, struct chfs_node_ref *,
51 struct chfs_inode *);
52 int chfs_gcollect_vnode(struct chfs_mount *, struct chfs_inode *);
53 int chfs_gcollect_dirent(struct chfs_mount *,
54 struct chfs_eraseblock *, struct chfs_inode *,
55 struct chfs_dirent *);
56 int chfs_gcollect_deletion_dirent(struct chfs_mount *,
57 struct chfs_eraseblock *, struct chfs_inode *,
58 struct chfs_dirent *);
59 int chfs_gcollect_dnode(struct chfs_mount *,
60 struct chfs_eraseblock *, struct chfs_inode *,
61 struct chfs_full_dnode *, uint32_t, uint32_t);
62
63 /* must be called with chm_lock_mountfields held */
64 void
65 chfs_gc_trigger(struct chfs_mount *chmp)
66 {
67 struct garbage_collector_thread *gc = &chmp->chm_gc_thread;
68
69 //mutex_enter(&chmp->chm_lock_sizes);
70 if (gc->gcth_running &&
71 chfs_gc_thread_should_wake(chmp)) {
72 cv_signal(&gc->gcth_wakeup);
73 }
74 //mutex_exit(&chmp->chm_lock_sizes);
75 }
76
77
78 void
79 chfs_gc_thread(void *data)
80 {
81 struct chfs_mount *chmp = data;
82 struct garbage_collector_thread *gc = &chmp->chm_gc_thread;
83
84 dbg_gc("[GC THREAD] thread started\n");
85
86 mutex_enter(&chmp->chm_lock_mountfields);
87 while (gc->gcth_running) {
88 /* we must call chfs_gc_thread_should_wake with chm_lock_mountfields
89 * held, which is a bit awkwardly done here, but we cant relly
90 * do it otherway with the current design...
91 */
92 if (chfs_gc_thread_should_wake(chmp)) {
93 // mutex_exit(&chmp->chm_lock_mountfields);
94 if (chfs_gcollect_pass(chmp) == ENOSPC) {
95 dbg_gc("No space for garbage collection\n");
96 panic("No space for garbage collection\n");
97 /* XXX why break here? i have added a panic
98 * here to see if it gets triggered -ahoka
99 */
100 break;
101 }
102 /* XXX gcollect_pass drops the mutex */
103 mutex_enter(&chmp->chm_lock_mountfields);
104 }
105
106 cv_timedwait_sig(&gc->gcth_wakeup,
107 &chmp->chm_lock_mountfields, mstohz(100));
108 }
109 mutex_exit(&chmp->chm_lock_mountfields);
110
111 dbg_gc("[GC THREAD] thread stopped\n");
112 kthread_exit(0);
113 }
114
115 void
116 chfs_gc_thread_start(struct chfs_mount *chmp)
117 {
118 struct garbage_collector_thread *gc = &chmp->chm_gc_thread;
119
120 cv_init(&gc->gcth_wakeup, "chfsgccv");
121
122 gc->gcth_running = true;
123 kthread_create(PRI_NONE, /*KTHREAD_MPSAFE |*/ KTHREAD_MUSTJOIN,
124 NULL, chfs_gc_thread, chmp, &gc->gcth_thread,
125 "chfsgcth");
126 }
127
128 void
129 chfs_gc_thread_stop(struct chfs_mount *chmp)
130 {
131 struct garbage_collector_thread *gc = &chmp->chm_gc_thread;
132
133 /* check if it is actually running. if not, do nothing */
134 if (gc->gcth_running) {
135 gc->gcth_running = false;
136 } else {
137 return;
138 }
139 cv_signal(&gc->gcth_wakeup);
140 dbg_gc("[GC THREAD] stop signal sent\n");
141
142 kthread_join(gc->gcth_thread);
143 #ifdef BROKEN_KTH_JOIN
144 kpause("chfsthjoin", false, mstohz(1000), NULL);
145 #endif
146
147 cv_destroy(&gc->gcth_wakeup);
148 }
149
150 /* must be called with chm_lock_mountfields held */
151 int
152 chfs_gc_thread_should_wake(struct chfs_mount *chmp)
153 {
154 int nr_very_dirty = 0;
155 struct chfs_eraseblock *cheb;
156 uint32_t dirty;
157
158 KASSERT(mutex_owned(&chmp->chm_lock_mountfields));
159
160 if (!TAILQ_EMPTY(&chmp->chm_erase_pending_queue)) {
161 dbg_gc("erase_pending\n");
162 return 1;
163 }
164
165 if (chmp->chm_unchecked_size) {
166 dbg_gc("unchecked\n");
167 return 1;
168 }
169
170 dirty = chmp->chm_dirty_size - chmp->chm_nr_erasable_blocks *
171 chmp->chm_ebh->eb_size;
172
173 if (chmp->chm_nr_free_blocks + chmp->chm_nr_erasable_blocks <
174 chmp->chm_resv_blocks_gctrigger && (dirty > chmp->chm_nospc_dirty)) {
175 dbg_gc("free: %d + erasable: %d < resv: %d\n",
176 chmp->chm_nr_free_blocks, chmp->chm_nr_erasable_blocks,
177 chmp->chm_resv_blocks_gctrigger);
178 dbg_gc("dirty: %d > nospc_dirty: %d\n",
179 dirty, chmp->chm_nospc_dirty);
180
181 return 1;
182 }
183
184 TAILQ_FOREACH(cheb, &chmp->chm_very_dirty_queue, queue) {
185 nr_very_dirty++;
186 if (nr_very_dirty == chmp->chm_vdirty_blocks_gctrigger) {
187 dbg_gc("nr_very_dirty\n");
188 return 1;
189 }
190 }
191
192 return 0;
193 }
194
195 void
196 chfs_gc_release_inode(struct chfs_mount *chmp,
197 struct chfs_inode *ip)
198 {
199 dbg_gc("release inode\n");
200 //mutex_exit(&ip->inode_lock);
201 //vput(ITOV(ip));
202 }
203
204 struct chfs_inode *
205 chfs_gc_fetch_inode(struct chfs_mount *chmp, ino_t vno,
206 uint32_t unlinked)
207 {
208 struct vnode *vp = NULL;
209 struct chfs_vnode_cache *vc;
210 struct chfs_inode *ip;
211 dbg_gc("fetch inode %llu\n", (unsigned long long)vno);
212
213 if (unlinked) {
214 dbg_gc("unlinked\n");
215 vp = chfs_vnode_lookup(chmp, vno);
216 if (!vp) {
217 mutex_enter(&chmp->chm_lock_vnocache);
218 vc = chfs_vnode_cache_get(chmp, vno);
219 if (!vc) {
220 mutex_exit(&chmp->chm_lock_vnocache);
221 return NULL;
222 }
223 if (vc->state != VNO_STATE_CHECKEDABSENT) {
224 //sleep_on_spinunlock(&chmp->chm_lock_vnocache);
225 mutex_exit(&chmp->chm_lock_vnocache);
226 /* XXX why do we need the delay here?! */
227 // kpause("chvncabs", true, mstohz(50), NULL);
228 KASSERT(mutex_owned(&chmp->chm_lock_mountfields));
229 cv_timedwait_sig(
230 &chmp->chm_gc_thread.gcth_wakeup,
231 &chmp->chm_lock_mountfields, mstohz(50));
232
233 // KASSERT(!mutex_owned(&chmp->chm_lock_vnocache));
234 } else {
235 mutex_exit(&chmp->chm_lock_vnocache);
236 }
237 return NULL;
238 }
239 } else {
240 dbg_gc("vnode lookup\n");
241 vp = chfs_vnode_lookup(chmp, vno);
242 //VFS_VGET(chmp->chm_fsmp, vno, &vp);
243 }
244 dbg_gc("vp to ip\n");
245 ip = VTOI(vp);
246 KASSERT(ip);
247 //mutex_enter(&ip->inode_lock);
248
249 return ip;
250 }
251
252 extern rb_tree_ops_t frag_rbtree_ops;
253
254 int
255 chfs_check(struct chfs_mount *chmp, struct chfs_vnode_cache *chvc)
256 {
257 struct chfs_inode *ip;
258 struct vnode *vp;
259 int ret;
260
261 ip = pool_get(&chfs_inode_pool, PR_WAITOK);
262 if (!ip) {
263 return ENOMEM;
264 }
265
266 vp = kmem_zalloc(sizeof(struct vnode), KM_SLEEP);
267
268 ip->chvc = chvc;
269 ip->vp = vp;
270
271 vp->v_data = ip;
272
273 rb_tree_init(&ip->fragtree, &frag_rbtree_ops);
274 TAILQ_INIT(&ip->dents);
275
276 ret = chfs_read_inode_internal(chmp, ip);
277 if (!ret) {
278 chfs_clear_inode(chmp, ip);
279 }
280
281 pool_put(&chfs_inode_pool, ip);
282
283 return ret;
284 }
285
286 void
287 chfs_clear_inode(struct chfs_mount *chmp, struct chfs_inode *ip)
288 {
289 struct chfs_dirent *fd, *tmpfd;
290 struct chfs_vnode_cache *chvc;
291
292
293 /* XXX not sure if this is the correct locking */
294 // mutex_enter(&chmp->chm_lock_vnocache);
295 chvc = ip->chvc;
296 /* shouldnt this be: */
297 //bool deleted = (chvc && !(chvc->pvno || chvc->nlink));
298 int deleted = (chvc && !(chvc->pvno | chvc->nlink));
299
300 if (chvc && chvc->state != VNO_STATE_CHECKING) {
301 // chfs_vnode_cache_state_set(chmp, chvc, VNO_STATE_CLEARING);
302 chvc->state = VNO_STATE_CLEARING;
303 }
304
305 if (chvc->v && ((struct chfs_vnode_cache *)chvc->v != chvc)) {
306 if (deleted)
307 chfs_mark_node_obsolete(chmp, chvc->v);
308 //chfs_free_refblock(chvc->v);
309 }
310 // mutex_enter(&chmp->chm_lock_vnocache);
311
312 chfs_kill_fragtree(&ip->fragtree);
313 /*
314 fd = TAILQ_FIRST(&ip->dents);
315 while (fd) {
316 TAILQ_REMOVE(&ip->dents, fd, fds);
317 chfs_free_dirent(fd);
318 fd = TAILQ_FIRST(&ip->dents);
319 }
320 */
321
322 TAILQ_FOREACH_SAFE(fd, &ip->dents, fds, tmpfd) {
323 chfs_free_dirent(fd);
324 }
325
326 if (chvc && chvc->state == VNO_STATE_CHECKING) {
327 chfs_vnode_cache_set_state(chmp,
328 chvc, VNO_STATE_CHECKEDABSENT);
329 if ((struct chfs_vnode_cache *)chvc->v == chvc &&
330 (struct chfs_vnode_cache *)chvc->dirents == chvc &&
331 (struct chfs_vnode_cache *)chvc->dnode == chvc)
332 chfs_vnode_cache_remove(chmp, chvc);
333 }
334
335 }
336
337 struct chfs_eraseblock *
338 find_gc_block(struct chfs_mount *chmp)
339 {
340 struct chfs_eraseblock *ret;
341 struct chfs_eraseblock_queue *nextqueue;
342
343 KASSERT(mutex_owned(&chmp->chm_lock_mountfields));
344
345 struct timespec now;
346 vfs_timestamp(&now);
347
348 int n = now.tv_nsec % 128;
349
350 //dbg_gc("n = %d\n", n);
351 again:
352 /* if (!TAILQ_EMPTY(&chmp->chm_bad_used_queue) && chmp->chm_nr_free_blocks > chmp->chm_nr_resv_blocks_gcbad) {
353 dbg_gc("Picking block from bad_used_queue to GC next\n");
354 nextqueue = &chmp->chm_bad_used_queue;
355 } else */if (n<50 && !TAILQ_EMPTY(&chmp->chm_erase_pending_queue)) {
356 dbg_gc("Picking block from erase_pending_queue to GC next\n");
357 nextqueue = &chmp->chm_erase_pending_queue;
358 } else if (n<110 && !TAILQ_EMPTY(&chmp->chm_very_dirty_queue) ) {
359 dbg_gc("Picking block from very_dirty_queue to GC next\n");
360 nextqueue = &chmp->chm_very_dirty_queue;
361 } else if (n<126 && !TAILQ_EMPTY(&chmp->chm_dirty_queue) ) {
362 dbg_gc("Picking block from dirty_queue to GC next\n");
363 nextqueue = &chmp->chm_dirty_queue;
364 } else if (!TAILQ_EMPTY(&chmp->chm_clean_queue)) {
365 dbg_gc("Picking block from clean_queue to GC next\n");
366 nextqueue = &chmp->chm_clean_queue;
367 } else if (!TAILQ_EMPTY(&chmp->chm_dirty_queue)) {
368 dbg_gc("Picking block from dirty_queue to GC next"
369 " (clean_queue was empty)\n");
370 nextqueue = &chmp->chm_dirty_queue;
371 } else if (!TAILQ_EMPTY(&chmp->chm_very_dirty_queue)) {
372 dbg_gc("Picking block from very_dirty_queue to GC next"
373 " (clean_queue and dirty_queue were empty)\n");
374 nextqueue = &chmp->chm_very_dirty_queue;
375 } else if (!TAILQ_EMPTY(&chmp->chm_erase_pending_queue)) {
376 dbg_gc("Picking block from erase_pending_queue to GC next"
377 " (clean_queue and {very_,}dirty_queue were empty)\n");
378 nextqueue = &chmp->chm_erase_pending_queue;
379 } else if (!TAILQ_EMPTY(&chmp->chm_erasable_pending_wbuf_queue)) {
380 dbg_gc("Synching wbuf in order to reuse "
381 "erasable_pendig_wbuf_queue blocks\n");
382 rw_enter(&chmp->chm_lock_wbuf, RW_WRITER);
383 chfs_flush_pending_wbuf(chmp);
384 rw_exit(&chmp->chm_lock_wbuf);
385 goto again;
386 } else {
387 dbg_gc("CHFS: no clean, dirty _or_ erasable"
388 " blocks to GC from! Where are they all?\n");
389 return NULL;
390 }
391
392 ret = TAILQ_FIRST(nextqueue);
393 if (chmp->chm_nextblock) {
394 dbg_gc("nextblock num: %u - gcblock num: %u\n",
395 chmp->chm_nextblock->lnr, ret->lnr);
396 if (ret == chmp->chm_nextblock)
397 goto again;
398 //KASSERT(ret != chmp->chm_nextblock);
399 //dbg_gc("first node lnr: %u ofs: %u\n", ret->first_node->lnr, ret->first_node->offset);
400 //dbg_gc("last node lnr: %u ofs: %u\n", ret->last_node->lnr, ret->last_node->offset);
401 }
402 TAILQ_REMOVE(nextqueue, ret, queue);
403 chmp->chm_gcblock = ret;
404 ret->gc_node = ret->first_node;
405
406 if (!ret->gc_node) {
407 dbg_gc("Oops! ret->gc_node at LEB: %u is NULL\n", ret->lnr);
408 panic("CHFS BUG - one LEB's gc_node is NULL\n");
409 }
410
411 /* TODO wasted size? */
412 return ret;
413 }
414
415
416 int
417 chfs_gcollect_pass(struct chfs_mount *chmp)
418 {
419 struct chfs_vnode_cache *vc;
420 struct chfs_eraseblock *eb;
421 struct chfs_node_ref *nref;
422 uint32_t gcblock_dirty;
423 struct chfs_inode *ip;
424 ino_t vno, pvno;
425 uint32_t nlink;
426 int ret = 0;
427
428 KASSERT(mutex_owned(&chmp->chm_lock_mountfields));
429
430 // mutex_enter(&chmp->chm_lock_mountfields);
431 for (;;) {
432 mutex_enter(&chmp->chm_lock_sizes);
433
434 dbg_gc("unchecked size == %u\n", chmp->chm_unchecked_size);
435 if (!chmp->chm_unchecked_size)
436 break;
437
438 if (chmp->chm_checked_vno > chmp->chm_max_vno) {
439 mutex_exit(&chmp->chm_lock_sizes);
440 mutex_exit(&chmp->chm_lock_mountfields);
441 dbg_gc("checked_vno (#%llu) > max_vno (#%llu)\n",
442 (unsigned long long)chmp->chm_checked_vno,
443 (unsigned long long)chmp->chm_max_vno);
444 return ENOSPC;
445 }
446
447 mutex_exit(&chmp->chm_lock_sizes);
448
449 mutex_enter(&chmp->chm_lock_vnocache);
450 dbg_gc("checking vno #%llu\n",
451 (unsigned long long)chmp->chm_checked_vno);
452 dbg_gc("get vnode cache\n");
453 vc = chfs_vnode_cache_get(chmp, chmp->chm_checked_vno++);
454
455 if (!vc) {
456 dbg_gc("!vc\n");
457 mutex_exit(&chmp->chm_lock_vnocache);
458 continue;
459 }
460
461 if ((vc->pvno | vc->nlink) == 0) {
462 dbg_gc("(pvno | nlink) == 0\n");
463 mutex_exit(&chmp->chm_lock_vnocache);
464 continue;
465 }
466
467 dbg_gc("switch\n");
468 switch (vc->state) {
469 case VNO_STATE_CHECKEDABSENT:
470 case VNO_STATE_PRESENT:
471 mutex_exit(&chmp->chm_lock_vnocache);
472 continue;
473
474 case VNO_STATE_GC:
475 case VNO_STATE_CHECKING:
476 mutex_exit(&chmp->chm_lock_vnocache);
477 mutex_exit(&chmp->chm_lock_mountfields);
478 dbg_gc("VNO_STATE GC or CHECKING\n");
479 panic("CHFS BUG - vc state gc or checking\n");
480
481 case VNO_STATE_READING:
482 chmp->chm_checked_vno--;
483 mutex_exit(&chmp->chm_lock_vnocache);
484 /* XXX why do we need the delay here?! */
485 kpause("chvncrea", true, mstohz(50), NULL);
486
487 // sleep_on_spinunlock(&chmp->chm_lock_vnocache);
488 // KASSERT(!mutex_owned(&chmp->chm_lock_vnocache));
489 mutex_exit(&chmp->chm_lock_mountfields);
490 return 0;
491
492 default:
493 mutex_exit(&chmp->chm_lock_vnocache);
494 mutex_exit(&chmp->chm_lock_mountfields);
495 dbg_gc("default\n");
496 panic("CHFS BUG - vc state is other what we"
497 " checked\n");
498
499 case VNO_STATE_UNCHECKED:
500 ;
501 }
502
503 chfs_vnode_cache_set_state(chmp, vc, VNO_STATE_CHECKING);
504
505 /* XXX check if this is too heavy to call under
506 * chm_lock_vnocache
507 */
508 ret = chfs_check(chmp, vc);
509 dbg_gc("set state\n");
510 chfs_vnode_cache_set_state(chmp,
511 vc, VNO_STATE_CHECKEDABSENT);
512
513 mutex_exit(&chmp->chm_lock_vnocache);
514 mutex_exit(&chmp->chm_lock_mountfields);
515
516 return ret;
517 }
518
519
520 eb = chmp->chm_gcblock;
521
522 if (!eb) {
523 eb = find_gc_block(chmp);
524 }
525
526 if (!eb) {
527 dbg_gc("!eb\n");
528 if (!TAILQ_EMPTY(&chmp->chm_erase_pending_queue)) {
529 mutex_exit(&chmp->chm_lock_sizes);
530 mutex_exit(&chmp->chm_lock_mountfields);
531 return EAGAIN;
532 }
533 mutex_exit(&chmp->chm_lock_sizes);
534 mutex_exit(&chmp->chm_lock_mountfields);
535 return EIO;
536 }
537
538 if (!eb->used_size) {
539 dbg_gc("!eb->used_size\n");
540 goto eraseit;
541 }
542
543 nref = eb->gc_node;
544 //dbg_gc("gc use: %u\n", chmp->chm_nextblock->lnr);
545 //dbg_gc("nref: %u %u\n", nref->nref_lnr, nref->nref_offset);
546 gcblock_dirty = eb->dirty_size;
547
548 while(CHFS_REF_OBSOLETE(nref)) {
549 //dbg_gc("obsoleted nref lnr: %u - offset: %u\n", nref->nref_lnr, nref->nref_offset);
550 #ifdef DBG_MSG_GC
551 if (nref == chmp->chm_blocks[nref->nref_lnr].last_node) {
552 dbg_gc("THIS NODE IS THE LAST NODE OF ITS EB\n");
553 }
554 #endif
555 nref = node_next(nref);
556 if (!nref) {
557 //dbg_gc("!nref\n");
558 eb->gc_node = nref;
559 mutex_exit(&chmp->chm_lock_sizes);
560 mutex_exit(&chmp->chm_lock_mountfields);
561 panic("CHFS BUG - nref is NULL)\n");
562 }
563 }
564 eb->gc_node = nref;
565 //dbg_gc("nref the chosen one lnr: %u - offset: %u\n", nref->nref_lnr, nref->nref_offset);
566 KASSERT(nref->nref_lnr == chmp->chm_gcblock->lnr);
567
568 if (!nref->nref_next) {
569 //dbg_gc("!nref->nref_next\n");
570 mutex_exit(&chmp->chm_lock_sizes);
571 if (CHFS_REF_FLAGS(nref) == CHFS_PRISTINE_NODE_MASK) {
572 chfs_gcollect_pristine(chmp, eb, NULL, nref);
573 } else {
574 chfs_mark_node_obsolete(chmp, nref);
575 }
576 goto lock_size;
577 }
578
579 dbg_gc("nref lnr: %u - offset: %u\n", nref->nref_lnr, nref->nref_offset);
580 vc = chfs_nref_to_vc(nref);
581
582 mutex_exit(&chmp->chm_lock_sizes);
583
584 //dbg_gc("enter vnocache lock on #%llu\n", vc->vno);
585 mutex_enter(&chmp->chm_lock_vnocache);
586
587 dbg_gc("switch\n");
588 switch(vc->state) {
589 case VNO_STATE_CHECKEDABSENT:
590 if (CHFS_REF_FLAGS(nref) == CHFS_PRISTINE_NODE_MASK) {
591 chfs_vnode_cache_set_state(chmp, vc, VNO_STATE_GC);
592 }
593 break;
594
595 case VNO_STATE_PRESENT:
596 break;
597
598 case VNO_STATE_UNCHECKED:
599 case VNO_STATE_CHECKING:
600 case VNO_STATE_GC:
601 mutex_exit(&chmp->chm_lock_vnocache);
602 mutex_exit(&chmp->chm_lock_mountfields);
603 panic("CHFS BUG - vc state unchecked,"
604 " checking or gc (vno #%llu, num #%d)\n",
605 (unsigned long long)vc->vno, vc->state);
606
607 case VNO_STATE_READING:
608 mutex_exit(&chmp->chm_lock_vnocache);
609 /* XXX why do we need the delay here?! */
610 kpause("chvncrea", true, mstohz(50), NULL);
611
612 // sleep_on_spinunlock(&chmp->chm_lock_vnocache);
613 // KASSERT(!mutex_owned(&chmp->chm_lock_vnocache));
614 mutex_exit(&chmp->chm_lock_mountfields);
615 return 0;
616 }
617
618 if (vc->state == VNO_STATE_GC) {
619 dbg_gc("vc->state == VNO_STATE_GC\n");
620 mutex_exit(&chmp->chm_lock_vnocache);
621 ret = chfs_gcollect_pristine(chmp, eb, NULL, nref);
622
623 // chfs_vnode_cache_state_set(chmp,
624 // vc, VNO_STATE_CHECKEDABSENT);
625 /* XXX locking? */
626 vc->state = VNO_STATE_CHECKEDABSENT;
627 //TODO wake_up(&chmp->chm_vnocache_wq);
628 if (ret != EBADF)
629 goto test_gcnode;
630 mutex_enter(&chmp->chm_lock_vnocache);
631 }
632
633 vno = vc->vno;
634 pvno = vc->pvno;
635 nlink = vc->nlink;
636 mutex_exit(&chmp->chm_lock_vnocache);
637
638 ip = chfs_gc_fetch_inode(chmp, vno, !(pvno | nlink));
639
640 if (!ip) {
641 dbg_gc("!ip\n");
642 ret = 0;
643 goto lock_size;
644 }
645
646 chfs_gcollect_live(chmp, eb, nref, ip);
647
648 chfs_gc_release_inode(chmp, ip);
649
650 test_gcnode:
651 if (eb->dirty_size == gcblock_dirty &&
652 !CHFS_REF_OBSOLETE(eb->gc_node)) {
653 dbg_gc("ERROR collecting node at %u failed.\n",
654 CHFS_GET_OFS(eb->gc_node->nref_offset));
655
656 ret = ENOSPC;
657 }
658
659 lock_size:
660 KASSERT(mutex_owned(&chmp->chm_lock_mountfields));
661 mutex_enter(&chmp->chm_lock_sizes);
662 eraseit:
663 dbg_gc("eraseit\n");
664
665 if (chmp->chm_gcblock) {
666 dbg_gc("eb used size = %u\n", chmp->chm_gcblock->used_size);
667 dbg_gc("eb free size = %u\n", chmp->chm_gcblock->free_size);
668 dbg_gc("eb dirty size = %u\n", chmp->chm_gcblock->dirty_size);
669 dbg_gc("eb unchecked size = %u\n",
670 chmp->chm_gcblock->unchecked_size);
671 dbg_gc("eb wasted size = %u\n", chmp->chm_gcblock->wasted_size);
672
673 KASSERT(chmp->chm_gcblock->used_size + chmp->chm_gcblock->free_size +
674 chmp->chm_gcblock->dirty_size +
675 chmp->chm_gcblock->unchecked_size +
676 chmp->chm_gcblock->wasted_size == chmp->chm_ebh->eb_size);
677
678 }
679
680 if (chmp->chm_gcblock && chmp->chm_gcblock->dirty_size +
681 chmp->chm_gcblock->wasted_size == chmp->chm_ebh->eb_size) {
682 dbg_gc("Block at leb #%u completely obsoleted by GC, "
683 "Moving to erase_pending_queue\n", chmp->chm_gcblock->lnr);
684 TAILQ_INSERT_TAIL(&chmp->chm_erase_pending_queue,
685 chmp->chm_gcblock, queue);
686 chmp->chm_gcblock = NULL;
687 chmp->chm_nr_erasable_blocks++;
688 if (!TAILQ_EMPTY(&chmp->chm_erase_pending_queue)) {
689 ret = chfs_remap_leb(chmp);
690 }
691 }
692
693 mutex_exit(&chmp->chm_lock_sizes);
694 mutex_exit(&chmp->chm_lock_mountfields);
695 dbg_gc("return\n");
696 return ret;
697 }
698
699
700 int
701 chfs_gcollect_pristine(struct chfs_mount *chmp, struct chfs_eraseblock *cheb,
702 struct chfs_vnode_cache *chvc, struct chfs_node_ref *nref)
703 {
704 struct chfs_node_ref *newnref;
705 struct chfs_flash_node_hdr *nhdr;
706 struct chfs_flash_vnode *fvnode;
707 struct chfs_flash_dirent_node *fdirent;
708 struct chfs_flash_data_node *fdata;
709 int ret, retries = 0;
710 uint32_t ofs, crc;
711 size_t totlen = chfs_nref_len(chmp, cheb, nref);
712 char *data;
713 struct iovec vec;
714 size_t retlen;
715
716 dbg_gc("gcollect_pristine\n");
717
718 data = kmem_alloc(totlen, KM_SLEEP);
719 if (!data)
720 return ENOMEM;
721
722 ofs = CHFS_GET_OFS(nref->nref_offset);
723
724 ret = chfs_read_leb(chmp, nref->nref_lnr, data, ofs, totlen, &retlen);
725 if (ret) {
726 dbg_gc("reading error\n");
727 return ret;
728 }
729 if (retlen != totlen) {
730 dbg_gc("read size error\n");
731 return EIO;
732 }
733 nhdr = (struct chfs_flash_node_hdr *)data;
734 /* check the header */
735 if (le16toh(nhdr->magic) != CHFS_FS_MAGIC_BITMASK) {
736 dbg_gc("node header magic number error\n");
737 return EBADF;
738 }
739 crc = crc32(0, (uint8_t *)nhdr, CHFS_NODE_HDR_SIZE - 4);
740 if (crc != le32toh(nhdr->hdr_crc)) {
741 dbg_gc("node header crc error\n");
742 return EBADF;
743 }
744
745 switch(le16toh(nhdr->type)) {
746 case CHFS_NODETYPE_VNODE:
747 fvnode = (struct chfs_flash_vnode *)data;
748 crc = crc32(0, (uint8_t *)fvnode, sizeof(struct chfs_flash_vnode) - 4);
749 if (crc != le32toh(fvnode->node_crc)) {
750 dbg_gc("vnode crc error\n");
751 return EBADF;
752 }
753 break;
754 case CHFS_NODETYPE_DIRENT:
755 fdirent = (struct chfs_flash_dirent_node *)data;
756 crc = crc32(0, (uint8_t *)fdirent, sizeof(struct chfs_flash_dirent_node) - 4);
757 if (crc != le32toh(fdirent->node_crc)) {
758 dbg_gc("dirent crc error\n");
759 return EBADF;
760 }
761 crc = crc32(0, fdirent->name, fdirent->nsize);
762 if (crc != le32toh(fdirent->name_crc)) {
763 dbg_gc("dirent name crc error\n");
764 return EBADF;
765 }
766 break;
767 case CHFS_NODETYPE_DATA:
768 fdata = (struct chfs_flash_data_node *)data;
769 crc = crc32(0, (uint8_t *)fdata, sizeof(struct chfs_flash_data_node) - 4);
770 if (crc != le32toh(fdata->node_crc)) {
771 dbg_gc("data node crc error\n");
772 return EBADF;
773 }
774 break;
775 default:
776 if (chvc) {
777 dbg_gc("unknown node have vnode cache\n");
778 return EBADF;
779 }
780 }
781 /* CRC's OK, write node to its new place */
782 retry:
783 ret = chfs_reserve_space_gc(chmp, totlen);
784 if (ret)
785 return ret;
786
787 newnref = chfs_alloc_node_ref(chmp->chm_nextblock);
788 if (!newnref)
789 return ENOMEM;
790
791 ofs = chmp->chm_ebh->eb_size - chmp->chm_nextblock->free_size;
792 newnref->nref_offset = ofs;
793
794 vec.iov_base = (void *)data;
795 vec.iov_len = totlen;
796 mutex_enter(&chmp->chm_lock_sizes);
797 ret = chfs_write_wbuf(chmp, &vec, 1, ofs, &retlen);
798
799 if (ret || retlen != totlen) {
800 chfs_err("error while writing out to the media\n");
801 chfs_err("err: %d | size: %zu | retlen : %zu\n",
802 ret, totlen, retlen);
803
804 chfs_change_size_dirty(chmp, chmp->chm_nextblock, totlen);
805 if (retries) {
806 mutex_exit(&chmp->chm_lock_sizes);
807 return EIO;
808 }
809
810 retries++;
811 mutex_exit(&chmp->chm_lock_sizes);
812 goto retry;
813 }
814
815 mutex_exit(&chmp->chm_lock_sizes);
816 //TODO should we set free_size?
817 chfs_mark_node_obsolete(chmp, nref);
818 chfs_add_vnode_ref_to_vc(chmp, chvc, newnref);
819 return 0;
820 }
821
822
823 int
824 chfs_gcollect_live(struct chfs_mount *chmp,
825 struct chfs_eraseblock *cheb, struct chfs_node_ref *nref,
826 struct chfs_inode *ip)
827 {
828 struct chfs_node_frag *frag;
829 struct chfs_full_dnode *fn = NULL;
830 int start = 0, end = 0, nrfrags = 0;
831 struct chfs_dirent *fd = NULL;
832 int ret = 0;
833 bool is_dirent;
834
835 dbg_gc("gcollect_live\n");
836
837 if (chmp->chm_gcblock != cheb) {
838 dbg_gc("GC block is no longer gcblock. Restart.\n");
839 goto upnout;
840 }
841
842 if (CHFS_REF_OBSOLETE(nref)) {
843 dbg_gc("node to be GC'd was obsoleted in the meantime.\n");
844 goto upnout;
845 }
846
847 /* It's a vnode? */
848 if (ip->chvc->v == nref) {
849 chfs_gcollect_vnode(chmp, ip);
850 goto upnout;
851 }
852
853 /* find fn */
854 dbg_gc("find full dnode\n");
855 for(frag = frag_first(&ip->fragtree);
856 frag; frag = frag_next(&ip->fragtree, frag)) {
857 if (frag->node && frag->node->nref == nref) {
858 fn = frag->node;
859 end = frag->ofs + frag->size;
860 if (!nrfrags++)
861 start = frag->ofs;
862 if (nrfrags == frag->node->frags)
863 break;
864 }
865 }
866
867 /* It's a pristine node, or dnode (or hole? XXX have we hole nodes?) */
868 if (fn) {
869 if (CHFS_REF_FLAGS(nref) == CHFS_PRISTINE_NODE_MASK) {
870 ret = chfs_gcollect_pristine(chmp,
871 cheb, ip->chvc, nref);
872 if (!ret) {
873 frag->node->nref = ip->chvc->v;
874 }
875 if (ret != EBADF)
876 goto upnout;
877 }
878 //ret = chfs_gcollect_hole(chmp, cheb, ip, fn, start, end);
879 ret = chfs_gcollect_dnode(chmp, cheb, ip, fn, start, end);
880 goto upnout;
881 }
882
883
884 /* It's a dirent? */
885 dbg_gc("find full dirent\n");
886 is_dirent = false;
887 TAILQ_FOREACH(fd, &ip->dents, fds) {
888 if (fd->nref == nref) {
889 is_dirent = true;
890 break;
891 }
892 }
893
894 if (is_dirent && fd->vno) {
895 ret = chfs_gcollect_dirent(chmp, cheb, ip, fd);
896 } else if (is_dirent) {
897 ret = chfs_gcollect_deletion_dirent(chmp, cheb, ip, fd);
898 } else {
899 dbg_gc("Nref at leb #%u offset 0x%08x wasn't in node list"
900 " for ino #%llu\n",
901 nref->nref_lnr, CHFS_GET_OFS(nref->nref_offset),
902 (unsigned long long)ip->ino);
903 if (CHFS_REF_OBSOLETE(nref)) {
904 dbg_gc("But it's obsolete so we don't mind"
905 " too much.\n");
906 }
907 }
908
909 upnout:
910 return ret;
911 }
912
913 int
914 chfs_gcollect_vnode(struct chfs_mount *chmp, struct chfs_inode *ip)
915 {
916 int ret;
917 dbg_gc("gcollect_vnode\n");
918
919 ret = chfs_write_flash_vnode(chmp, ip, ALLOC_GC);
920
921 return ret;
922 }
923
924 int
925 chfs_gcollect_dirent(struct chfs_mount *chmp,
926 struct chfs_eraseblock *cheb, struct chfs_inode *parent,
927 struct chfs_dirent *fd)
928 {
929 struct vnode *vnode = NULL;
930 struct chfs_inode *ip;
931 struct chfs_node_ref *prev;
932 dbg_gc("gcollect_dirent\n");
933
934 vnode = chfs_vnode_lookup(chmp, fd->vno);
935
936 /* XXX maybe KASSERT or panic on this? */
937 if (vnode == NULL) {
938 return ENOENT;
939 }
940
941 ip = VTOI(vnode);
942
943 prev = parent->chvc->dirents;
944 if (prev == fd->nref) {
945 parent->chvc->dirents = prev->nref_next;
946 dbg_gc("fd nref removed from dirents list\n");
947 prev = NULL;
948 }
949 while (prev) {
950 if (prev->nref_next == fd->nref) {
951 prev->nref_next = fd->nref->nref_next;
952 dbg_gc("fd nref removed from dirents list\n");
953 break;
954 }
955 prev = prev->nref_next;
956 }
957
958 prev = fd->nref;
959 chfs_mark_node_obsolete(chmp, fd->nref);
960 return chfs_write_flash_dirent(chmp,
961 parent, ip, fd, fd->vno, ALLOC_GC);
962 }
963
964 /* Check dirents what are marked as deleted. */
965 int
966 chfs_gcollect_deletion_dirent(struct chfs_mount *chmp,
967 struct chfs_eraseblock *cheb, struct chfs_inode *parent,
968 struct chfs_dirent *fd)
969 {
970 struct chfs_flash_dirent_node chfdn;
971 struct chfs_node_ref *nref;
972 size_t retlen, name_len, nref_len;
973 uint32_t name_crc;
974
975 int ret;
976
977 struct vnode *vnode = NULL;
978
979 dbg_gc("gcollect_deletion_dirent\n");
980
981 name_len = strlen(fd->name);
982 name_crc = crc32(0, fd->name, name_len);
983
984 nref_len = chfs_nref_len(chmp, cheb, fd->nref);
985
986 vnode = chfs_vnode_lookup(chmp, fd->vno);
987
988 //dbg_gc("ip from vnode\n");
989 //VFS_VGET(chmp->chm_fsmp, fd->vno, &vnode);
990 //ip = VTOI(vnode);
991 //vput(vnode);
992
993 //dbg_gc("mutex enter erase_completion_lock\n");
994
995 // dbg_gc("alloc chfdn\n");
996 // chfdn = kmem_alloc(nref_len, KM_SLEEP);
997 // if (!chfdn)
998 // return ENOMEM;
999
1000 for (nref = parent->chvc->dirents;
1001 nref != (void*)parent->chvc;
1002 nref = nref->nref_next) {
1003
1004 if (!CHFS_REF_OBSOLETE(nref))
1005 continue;
1006
1007 /* if node refs have different length, skip */
1008 if (chfs_nref_len(chmp, NULL, nref) != nref_len)
1009 continue;
1010
1011 if (CHFS_GET_OFS(nref->nref_offset) ==
1012 CHFS_GET_OFS(fd->nref->nref_offset)) {
1013 continue;
1014 }
1015
1016 ret = chfs_read_leb(chmp,
1017 nref->nref_lnr, (void*)&chfdn, CHFS_GET_OFS(nref->nref_offset),
1018 nref_len, &retlen);
1019
1020 if (ret) {
1021 dbg_gc("Read error: %d\n", ret);
1022 continue;
1023 }
1024
1025 if (retlen != nref_len) {
1026 dbg_gc("Error reading node:"
1027 " read: %zu insted of: %zu\n", retlen, nref_len);
1028 continue;
1029 }
1030
1031 /* if node type doesn't match, skip */
1032 if (le16toh(chfdn.type) != CHFS_NODETYPE_DIRENT)
1033 continue;
1034
1035 /* if crc doesn't match, skip */
1036 if (le32toh(chfdn.name_crc) != name_crc)
1037 continue;
1038
1039 /* if length of name different, or this is an another deletion
1040 * dirent, skip
1041 */
1042 if (chfdn.nsize != name_len || !le64toh(chfdn.vno))
1043 continue;
1044
1045 /* check actual name */
1046 if (memcmp(chfdn.name, fd->name, name_len))
1047 continue;
1048
1049 // kmem_free(chfdn, nref_len);
1050
1051 chfs_mark_node_obsolete(chmp, fd->nref);
1052 return chfs_write_flash_dirent(chmp,
1053 parent, NULL, fd, fd->vno, ALLOC_GC);
1054 }
1055
1056 // kmem_free(chfdn, nref_len);
1057
1058 TAILQ_REMOVE(&parent->dents, fd, fds);
1059 chfs_free_dirent(fd);
1060 return 0;
1061 }
1062
1063 int
1064 chfs_gcollect_dnode(struct chfs_mount *chmp,
1065 struct chfs_eraseblock *orig_cheb, struct chfs_inode *ip,
1066 struct chfs_full_dnode *fn, uint32_t orig_start, uint32_t orig_end)
1067 {
1068 struct chfs_node_ref *nref, *prev;
1069 struct chfs_full_dnode *newfn;
1070 struct chfs_flash_data_node *fdnode;
1071 int ret = 0, retries = 0;
1072 uint32_t totlen;
1073 char *data = NULL;
1074 struct iovec vec;
1075 size_t retlen;
1076 dbg_gc("gcollect_dnode\n");
1077
1078 //uint32_t used_size;
1079
1080 /* TODO GC merging frags, should we use it?
1081
1082 uint32_t start, end;
1083
1084 start = orig_start;
1085 end = orig_end;
1086
1087 if (chmp->chm_nr_free_blocks + chmp->chm_nr_erasable_blocks > chmp->chm_resv_blocks_gcmerge) {
1088 struct chfs_node_frag *frag;
1089 uint32_t min, max;
1090
1091 min = start & (PAGE_CACHE_SIZE-1);
1092 max = min + PAGE_CACHE_SIZE;
1093
1094 frag = (struct chfs_node_frag *)rb_tree_find_node_leq(&ip->i_chfs_ext.fragtree, &start);
1095 KASSERT(frag->ofs == start);
1096
1097 while ((frag = frag_prev(&ip->i_chfs_ext.fragtree, frag)) && frag->ofs >= min) {
1098 if (frag->ofs > min) {
1099 start = frag->ofs;
1100 continue;
1101 }
1102
1103 if (!frag->node || !frag->node->nref) {
1104 break;
1105 } else {
1106 struct chfs_node_ref *nref = frag->node->nref;
1107 struct chfs_eraseblock *cheb;
1108
1109 cheb = &chmp->chm_blocks[nref->nref_lnr];
1110
1111 if (cheb == chmp->chm_gcblock)
1112 start = frag->ofs;
1113
1114 //TODO is this a clean block?
1115
1116 start = frag->ofs;
1117 break;
1118 }
1119 }
1120
1121 end--;
1122 frag = (struct chfs_node_frag *)rb_tree_find_node_leq(&ip->i_chfs_ext.fragtree, &(end));
1123
1124 while ((frag = frag_next(&ip->i_chfs_ext.fragtree, frag)) && (frag->ofs + frag->size <= max)) {
1125 if (frag->ofs + frag->size < max) {
1126 end = frag->ofs + frag->size;
1127 continue;
1128 }
1129
1130 if (!frag->node || !frag->node->nref) {
1131 break;
1132 } else {
1133 struct chfs_node_ref *nref = frag->node->nref;
1134 struct chfs_eraseblock *cheb;
1135
1136 cheb = &chmp->chm_blocks[nref->nref_lnr];
1137
1138 if (cheb == chmp->chm_gcblock)
1139 end = frag->ofs + frag->size;
1140
1141 //TODO is this a clean block?
1142
1143 end = frag->ofs + frag->size;
1144 break;
1145 }
1146 }
1147
1148 KASSERT(end <=
1149 frag_last(&ip->i_chfs_ext.fragtree)->ofs +
1150 frag_last(&ip->i_chfs_ext.fragtree)->size);
1151 KASSERT(end >= orig_end);
1152 KASSERT(start <= orig_start);
1153 }
1154 */
1155 KASSERT(orig_cheb->lnr == fn->nref->nref_lnr);
1156 totlen = chfs_nref_len(chmp, orig_cheb, fn->nref);
1157 data = kmem_alloc(totlen, KM_SLEEP);
1158
1159 ret = chfs_read_leb(chmp, fn->nref->nref_lnr, data, fn->nref->nref_offset,
1160 totlen, &retlen);
1161
1162 fdnode = (struct chfs_flash_data_node *)data;
1163 fdnode->version = htole64(++ip->chvc->highest_version);
1164 fdnode->node_crc = htole32(crc32(0, (uint8_t *)fdnode,
1165 sizeof(*fdnode) - 4));
1166
1167 vec.iov_base = (void *)data;
1168 vec.iov_len = totlen;
1169
1170 retry:
1171 ret = chfs_reserve_space_gc(chmp, totlen);
1172 if (ret)
1173 goto out;
1174
1175 nref = chfs_alloc_node_ref(chmp->chm_nextblock);
1176 if (!nref) {
1177 ret = ENOMEM;
1178 goto out;
1179 }
1180
1181 mutex_enter(&chmp->chm_lock_sizes);
1182
1183 nref->nref_offset = chmp->chm_ebh->eb_size - chmp->chm_nextblock->free_size;
1184 KASSERT(nref->nref_offset % 4 == 0);
1185 chfs_change_size_free(chmp, chmp->chm_nextblock, -totlen);
1186
1187 ret = chfs_write_wbuf(chmp, &vec, 1, nref->nref_offset, &retlen);
1188 if (ret || retlen != totlen) {
1189 chfs_err("error while writing out to the media\n");
1190 chfs_err("err: %d | size: %d | retlen : %zu\n",
1191 ret, totlen, retlen);
1192 chfs_change_size_dirty(chmp, chmp->chm_nextblock, totlen);
1193 if (retries) {
1194 ret = EIO;
1195 mutex_exit(&chmp->chm_lock_sizes);
1196 goto out;
1197 }
1198
1199 retries++;
1200 mutex_exit(&chmp->chm_lock_sizes);
1201 goto retry;
1202 }
1203
1204 dbg_gc("new nref lnr: %u - offset: %u\n", nref->nref_lnr, nref->nref_offset);
1205
1206 chfs_change_size_used(chmp, &chmp->chm_blocks[nref->nref_lnr], totlen);
1207 mutex_exit(&chmp->chm_lock_sizes);
1208 KASSERT(chmp->chm_blocks[nref->nref_lnr].used_size <= chmp->chm_ebh->eb_size);
1209
1210 newfn = chfs_alloc_full_dnode();
1211 newfn->nref = nref;
1212 newfn->ofs = fn->ofs;
1213 newfn->size = fn->size;
1214 newfn->frags = fn->frags;
1215
1216 //TODO should we remove fd from dnode list?
1217
1218 prev = ip->chvc->dnode;
1219 if (prev == fn->nref) {
1220 ip->chvc->dnode = prev->nref_next;
1221 prev = NULL;
1222 }
1223 while (prev) {
1224 if (prev->nref_next == fn->nref) {
1225 prev->nref_next = fn->nref->nref_next;
1226 break;
1227 }
1228 prev = prev->nref_next;
1229 }
1230
1231 chfs_add_full_dnode_to_inode(chmp, ip, newfn);
1232 chfs_add_node_to_list(chmp,
1233 ip->chvc, newfn->nref, &ip->chvc->dnode);
1234
1235 out:
1236 kmem_free(data, totlen);
1237 return ret;
1238 }
1239