Lines Matching refs:first

42  * The vm_getusage() calculation is implemented so that the first thread
665 end, char type, vmu_bound_t **first, vmu_bound_t **last)
674 *first = *last = NULL;
684 *first = walker;
694 *first = walker;
700 * At this point, if *first is still NULL, then we
708 if (*first == NULL) {
713 *first = tmp;
717 *first = *last = tmp;
723 ASSERT(*first != NULL);
734 *last = *first;
773 * tree: avl_tree in which first and last hang.
775 * first, last: list of continuous bounds, of which zero or more are of
782 * update the types of bounds in (first,last) with
785 * For the list of bounds (first,last), this function updates any bounds
793 * The number of pages in the list of bounds (first,last) that were of
799 vmu_update_bounds(avl_tree_t *tree, vmu_bound_t **first, vmu_bound_t **last,
805 next = *first;
809 * Verify first and last bound are covered by new bounds if they
812 ASSERT((*first)->vmb_type != VMUSAGE_BOUND_UNKNOWN ||
813 (*first)->vmb_start >= new_first->vmb_start);
853 * Merges adjacent bounds with same type between first and last bound.
858 vmu_merge_bounds(avl_tree_t *tree, vmu_bound_t **first, vmu_bound_t **last)
864 ASSERT(*first != NULL);
867 current = *first;
889 * first and last may be modified, as bounds may be split into multiple
897 vmu_bound_t **first, vmu_bound_t **last, boolean_t incore)
908 next = *first;
992 vmu_bound_t **first, vmu_bound_t **last)
1000 next = *first;
1079 * For private mapped files, first the amp is checked for private pages.
1093 vmu_bound_t *first = NULL;
1241 if (first == NULL) {
1242 first = cur;
1351 if (first == NULL) {
1357 first = vmu_alloc_bound();
1358 first->vmb_start = s_start;
1359 first->vmb_end = s_end;
1360 first->vmb_type = VMUSAGE_BOUND_UNKNOWN;
1366 cur = first;
1371 &first, &last) > 0) {
1378 shared_object->vmo_key, &first,
1384 shared_object->vmo_key, &first,
1387 vmu_merge_bounds(tree, &first, &last);
1413 &e_last, tree, first, last);