Lines Matching defs:lib_va
230 struct lib_va {
231 struct lib_va *lv_next;
253 static struct lib_va *lib_va_hash[LIB_VA_SIZE];
276 * lib_va will be used for optimized allocation of address ranges for
294 * Number of 32 bit and 64 bit libraries in lib_va hash.
301 * We also decrement the number of libraries mapped via a lib_va
305 lib_va_free(struct lib_va *lvp)
319 kmem_free(lvp, sizeof (struct lib_va));
323 * See if the file associated with the vap passed in is in the lib_va hash.
328 static struct lib_va *
331 struct lib_va *lvp;
332 struct lib_va *del = NULL;
333 struct lib_va **tmp;
379 * Add a new entry to the lib_va hash.
392 static struct lib_va *
395 struct lib_va *lvp;
398 struct lib_va **tmp;
399 struct lib_va *del = NULL;
404 lvp = kmem_alloc(sizeof (struct lib_va), KM_SLEEP);
418 kmem_free(lvp, sizeof (struct lib_va));
500 lib_va_release(struct lib_va *lvp)
648 mmapobj_lookup_start_addr(struct lib_va *lvp)
706 * to the caller. If we're using lib_va and we need to allocate an address,
713 * into the lib_va hash, then *lvpp will be set to point to this lib_va
721 mmapobj_alloc_start_addr(struct lib_va **lvpp, size_t len, int use_lib_va,
759 * The first time through, we need to setup the lib_va arenas.
764 * for lib_va addresses but we do not prevent any allocations
852 * NULL instead of looking it up in the lib_va hash.
1559 struct lib_va *lvp = NULL;
1695 * have it if we found all the info we need in the lib_va cache.
1719 * the start address since we were not going to use lib_va.
1720 * Since we know that lib_va will not be used, it's safe
2318 struct lib_va *lvp;
2339 * Check lib_va to see if we already have a full description