Lines Matching refs:am_anon
198 amap->am_anon = kmem_alloc(totalslots * sizeof(struct vm_anon *),
200 if (amap->am_anon == NULL)
245 memset(amap->am_anon, 0,
332 kmem_free(amap->am_anon, slots * sizeof(*amap->am_anon));
415 KASSERT(amap->am_anon[i] == NULL);
418 KASSERT(amap->am_anon[i] == NULL);
480 * no need to zero am_anon since that was done at
512 memmove(amap->am_anon + slotarea,
513 amap->am_anon + slotoff,
515 memset(amap->am_anon + slotoff, 0,
610 /* do am_anon */
611 oldover = amap->am_anon;
623 amap->am_anon = newover;
698 * all mappings of a shared amap.] we traverse am_anon or am_slots
717 * Cheaper to traverse am_anon.
720 anon = amap->am_anon[lcv];
739 anon = amap->am_anon[slot];
777 anon = amap->am_anon[slot];
958 amap->am_anon[lcv] =
959 srcamap->am_anon[entry->aref.ar_pageoff + lcv];
960 if (amap->am_anon[lcv] == NULL)
962 KASSERT(amap->am_anon[lcv]->an_lock == srcamap->am_lock);
963 KASSERT(amap->am_anon[lcv]->an_ref > 0);
965 amap->am_anon[lcv]->an_ref++;
970 memset(&amap->am_anon[lcv], 0,
1036 * am_anon[] array on us while the lock is dropped.
1043 anon = amap->am_anon[slot];
1119 amap->am_anon[slot] = nanon;
1278 * We can either traverse the amap by am_anon or by am_slots.
1300 if (amap->am_anon[curslot] == NULL)
1310 anon = amap->am_anon[curslot];
1317 amap->am_anon[curslot] = NULL;
1395 anon = am->am_anon[slot];
1456 an = amap->am_anon[slot];
1490 memcpy(anons, &amap->am_anon[slot], npages * sizeof(struct vm_anon *));
1527 struct vm_anon *oanon = amap->am_anon[slot];
1537 KASSERT(amap->am_anon[slot] == NULL);
1543 amap->am_anon[slot] = anon;
1566 KASSERT(amap->am_anon[slot] != NULL);
1567 KASSERT(amap->am_anon[slot]->an_lock == amap->am_lock);
1569 amap->am_anon[slot] = NULL;