Lines Matching refs:len

77 static void segspt_free_pages(struct seg *seg, caddr_t addr, size_t len);
120 caddr_t addr, size_t len, enum fault_type type, enum seg_rw rw);
123 register size_t len, register uint_t prot);
128 static size_t segspt_shmincore(struct seg *seg, caddr_t addr, size_t len,
130 static int segspt_shmsync(struct seg *seg, register caddr_t addr, size_t len,
132 static int segspt_shmlockop(struct seg *seg, caddr_t addr, size_t len,
134 static int segspt_shmgetprot(struct seg *seg, caddr_t addr, size_t len,
139 static int segspt_shmadvise(struct seg *seg, caddr_t addr, size_t len,
179 static int spt_anon_getpages(struct seg *seg, caddr_t addr, size_t len,
264 segspt_shmsync(struct seg *seg, caddr_t addr, size_t len, int attr,
274 segspt_shmincore(struct seg *seg, caddr_t addr, size_t len, char *vec)
290 eo_seg = addr + len;
297 return (len);
311 npages = btopr(len);
337 return (len);
402 tnf_opaque, addr, addr, tnf_ulong, len, seg->s_size);
615 segspt_free_pages(struct seg *seg, caddr_t addr, size_t len)
635 len = P2ROUNDUP(len, PAGESIZE);
637 npages = btop(len);
645 hat_unload(seg->s_as->a_hat, addr, len, hat_flags);
824 segspt_dismpagelock(struct seg *seg, caddr_t addr, size_t len,
854 npages = btopr(len);
1180 segspt_shmpagelock(struct seg *seg, caddr_t addr, size_t len,
1210 return (segspt_dismpagelock(seg, addr, len, ppp, type, rw));
1214 npages = btopr(len);
1435 segspt_reclaim(void *ptag, caddr_t addr, size_t len, struct page **pplist,
1450 npages = (len >> PAGESHIFT);
1575 size_t len, enum seg_rw rw)
1614 hat_unlock(sptseg->s_as->a_hat, sptseg_addr, len);
1621 for (adr = sptseg_addr; adr < sptseg_addr + len; adr += PAGESIZE) {
1649 npages = btopr(len);
1804 segspt_shmsetprot(struct seg *seg, caddr_t addr, size_t len, uint_t prot)
1822 size_t len, enum fault_type type, enum seg_rw rw)
1857 if (((addr + len) - seg->s_base) > sptd->spt_realsize)
1861 * make any necessary adjustments to addr and len
1864 * The code in shmat() forces base addr and len of ISM segment
1875 size = P2ROUNDUP((uintptr_t)(((addr + len) - shm_addr)), pgsz);
2005 size_t len, enum fault_type type, enum seg_rw rw)
2034 return (segspt_dismfault(hat, seg, addr, len, type, rw));
2048 if (((addr + len) - seg->s_base) > sptd->spt_realsize)
2052 * make any necessary adjustments to addr and len
2055 * The code in shmat() forces base addr and len of ISM segment
2066 size = P2ROUNDUP((uintptr_t)(((addr + len) - shm_addr)), pgsz);
2325 size_t len,
2345 ASSERT(IS_P2ALIGNED(sptaddr, share_sz) && IS_P2ALIGNED(len, share_sz));
2346 ASSERT(len != 0);
2351 e_sptaddr = sptaddr + len;
2640 segspt_shmlockop(struct seg *seg, caddr_t addr, size_t len,
2651 size_t a_len; /* aligned len */
2669 npages = btopr(len);
2692 a_len = P2ROUNDUP((uintptr_t)(((addr + len) - a_addr)),
2784 segspt_shmgetprot(struct seg *seg, caddr_t addr, size_t len, uint_t *protv)
2788 spgcnt_t pgno = seg_page(seg, addr+len) - seg_page(seg, addr) + 1;
2850 segspt_shmadvise(struct seg *seg, caddr_t addr, size_t len, uint_t behav)
2874 (void) anon_disclaim(amp, pg_idx, len, behav, NULL);
2932 (void) anon_disclaim(amp, pg_idx, len, behav, NULL);
2950 size = P2ROUNDUP((uintptr_t)(((addr + len) - shm_addr)),
2970 policy = lgrp_madv_to_policy(behav, len, MAP_SHARED);
2972 NULL, 0, len);
3005 segspt_shmsetpgsz(struct seg *seg, caddr_t addr, size_t len, uint_t szc)