Lines Matching defs:chmp
62 struct chfs_mount* chmp;
99 chmp = ump->um_chfs;
101 ip->ino = ++chmp->chm_max_vno;
103 mutex_enter(&chmp->chm_lock_vnocache);
104 chvc = chfs_vnode_cache_get(chmp, ip->ino);
105 mutex_exit(&chmp->chm_lock_vnocache);
225 struct chfs_mount *chmp;
233 chmp = ump->um_chfs;
287 mutex_enter(&chmp->chm_lock_mountfields);
289 err = chfs_write_flash_dnode(chmp, vp, bp, fd);
291 mutex_exit(&chmp->chm_lock_mountfields);
297 err = chfs_add_full_dnode_to_inode(chmp, ip, fd);
299 mutex_exit(&chmp->chm_lock_mountfields);
304 mutex_exit(&chmp->chm_lock_mountfields);
460 struct chfs_mount *chmp = ump->um_chfs;
527 mutex_enter(&chmp->chm_lock_mountfields);
528 error = chfs_write_flash_vnode(chmp, ip, ALLOC_NORMAL);
529 mutex_exit(&chmp->chm_lock_mountfields);
586 /* calculates ((off_t)blk * chmp->chm_chm_fs_bsize) */
587 #define chfs_lblktosize(chmp, blk) \
588 (((off_t)(blk)) << (chmp)->chm_fs_bshift)
590 /* calculates (loc % chmp->chm_chm_fs_bsize) */
591 #define chfs_blkoff(chmp, loc) \
592 ((loc) & (chmp)->chm_fs_qbmask)
594 /* calculates (loc / chmp->chm_chm_fs_bsize) */
595 #define chfs_lblkno(chmp, loc) \
596 ((loc) >> (chmp)->chm_fs_bshift)
598 /* calculates roundup(size, chmp->chm_chm_fs_fsize) */
599 #define chfs_fragroundup(chmp, size) \
600 (((size) + (chmp)->chm_fs_qfmask) & (chmp)->chm_fs_fmask)
602 #define chfs_blksize(chmp, ip, lbn) \
603 (((lbn) >= UFS_NDADDR || (ip)->size >= chfs_lblktosize(chmp, (lbn) + 1)) \
604 ? (chmp)->chm_fs_bsize \
605 : (chfs_fragroundup(chmp, chfs_blkoff(chmp, (ip)->size))))
607 /* calculates roundup(size, chmp->chm_chm_fs_bsize) */
608 #define chfs_blkroundup(chmp, size) \
609 (((size) + (chmp)->chm_fs_qbmask) & (chmp)->chm_fs_bmask)
626 struct chfs_mount *chmp;
655 chmp = ip->chmp;
691 lbn = chfs_lblkno(chmp, uio->uio_offset);
693 size = chfs_blksize(chmp, ip, lbn);
694 blkoffset = chfs_blkoff(chmp, uio->uio_offset);
695 xfersize = MIN(MIN(chmp->chm_fs_bsize - blkoffset, uio->uio_resid),
698 if (chfs_lblktosize(chmp, nextlbn) >= ip->size) {
702 int nextsize = chfs_blksize(chmp, ip, nextlbn);
769 struct chfs_mount *chmp;
808 chmp = ip->chmp;
842 preallocoff = round_page(chfs_blkroundup(chmp,
846 endallocoff = nsize - chfs_blkoff(chmp, nsize);
853 if (nsize > osize && chfs_lblkno(chmp, osize) < UFS_NDADDR &&
854 chfs_lblkno(chmp, osize) != chfs_lblkno(chmp, nsize) &&
855 chfs_blkroundup(chmp, osize) != osize) {
858 eob = chfs_blkroundup(chmp, osize);
866 trunc_page(osize & chmp->chm_fs_bmask),
882 blkoffset = chfs_blkoff(chmp, uio->uio_offset);
883 bytelen = MIN(chmp->chm_fs_bsize - blkoffset, uio->uio_resid);
898 chfs_blkoff(chmp, uio->uio_offset) == 0 &&
903 len -= chfs_blkoff(chmp, len);
970 trunc_page(origoff & chmp->chm_fs_bmask),
971 round_page(chfs_blkroundup(chmp, uio->uio_offset)),
1001 mutex_enter(&chmp->chm_lock_mountfields);
1002 error = chfs_write_flash_vnode(chmp, ip, ALLOC_NORMAL);
1003 mutex_exit(&chmp->chm_lock_mountfields);
1268 struct chfs_mount *chmp;
1277 chmp = ump->um_chfs;
1301 mutex_enter(&chmp->chm_lock_mountfields);
1304 err = chfs_write_flash_dnode(chmp, vp, bp, fd);
1306 mutex_exit(&chmp->chm_lock_mountfields);
1311 err = chfs_add_full_dnode_to_inode(chmp, ip, fd);
1313 mutex_exit(&chmp->chm_lock_mountfields);
1317 mutex_exit(&chmp->chm_lock_mountfields);
1350 struct chfs_mount *chmp;
1384 chmp = ump->um_chfs;
1385 mutex_enter(&chmp->chm_lock_vnocache);
1386 chvc = chfs_vnode_cache_get(chmp, ip->ino);
1387 mutex_exit(&chmp->chm_lock_vnocache);
1502 struct chfs_mount *chmp = ip->chmp;
1507 mutex_enter(&chmp->chm_lock_mountfields);
1509 mutex_enter(&chmp->chm_lock_vnocache);
1511 mutex_exit(&chmp->chm_lock_vnocache);
1516 chfs_kill_fragtree(chmp, &ip->fragtree);
1536 mutex_exit(&chmp->chm_lock_mountfields);
1562 struct chfs_mount *chmp = ip->chmp;
1567 err = chfs_read_data(chmp, vp, bp);
1569 mutex_enter(&chmp->chm_lock_mountfields);
1573 err = chfs_write_flash_dnode(chmp, vp, bp, fd);
1575 mutex_exit(&chmp->chm_lock_mountfields);
1580 err = chfs_add_full_dnode_to_inode(chmp, ip, fd);
1582 mutex_exit(&chmp->chm_lock_mountfields);