/illumos-gate/usr/src/cmd/lp/filter/postscript/common/ |
H A D | request.h | 31 * or on a per page basis. All the translators I've supplied accept the -R option 34 * that can be "request", "request:page", or "request:page:file". If page isn't 36 * to the global environment, otherwise request holds only for the named page. 37 * If a file name is given a page number must be supplied, and in that case the 47 int page; member in struct:__anon794
|
H A D | request.c | 36 * page basis. Requests are passed through to the translator using the -R option. 37 * The argument to -R can be "request", "request:page", or "request:page:file". 38 * If page is omitted (as in the first form) or set to 0 request will be applied 40 * page. If a file is given, page must be supplied, and the lookup is in that file 53 Request request[MAXREQUEST]; /* next page or global request */ 66 char *page; /* and save it for this page */ local 71 * Save the request until we get to appropriate page 98 writerequest(int page, FILE *fp_out) argument [all...] |
/illumos-gate/usr/src/cmd/fm/modules/sun4/cpumem-diagnosis/ |
H A D | cmd_page.c | 29 * Support routines for managing per-page state. 45 page_write(fmd_hdl_t *hdl, cmd_page_t *page) argument 47 fmd_buf_write(hdl, NULL, page->page_bufname, page, 52 cmd_page_free(fmd_hdl_t *hdl, cmd_page_t *page, int destroy) argument 54 cmd_case_t *cc = &page->page_case; 66 fmd_buf_destroy(hdl, NULL, page->page_bufname); 68 cmd_fmri_fini(hdl, &page->page_asru, destroy); 70 cmd_list_delete(&cmd.cmd_pages, page); 71 fmd_hdl_free(hdl, page, sizeo 75 cmd_page_destroy(fmd_hdl_t *hdl, cmd_page_t *page) argument 83 cmd_page_t *page; local 97 cmd_page_t *page; local 161 cmd_page_t *page; local 177 cmd_page_t *page; local 276 cmd_page_t *page, *next; local 293 cmd_page_dirty(fmd_hdl_t *hdl, cmd_page_t *page) argument 307 cmd_page_t *page; local [all...] |
H A D | cmd_pageerr.c | 27 * Fault-handling routines for page retirement faults 47 cmd_page_t *page = NULL; local 54 page = cmd_page_lookup(afar); 55 if (page != NULL) { 57 * If the page has already been retired then *page 59 * flag would be 0x0 - check to see if the page 62 if (page->page_flags & CMD_MEM_F_FAULTING || 63 fmd_nvl_fmri_unusable(hdl, page->page_asru_nvl)) { 65 page [all...] |
/illumos-gate/usr/src/cmd/fm/modules/sun4v/generic-mem/ |
H A D | gmem_page.c | 27 * Support routines for managing per-page state. 41 page_write(fmd_hdl_t *hdl, gmem_page_t *page) argument 43 fmd_buf_write(hdl, NULL, page->page_bufname, page, 48 gmem_page_free(fmd_hdl_t *hdl, gmem_page_t *page, int destroy) argument 50 gmem_case_t *cc = &page->page_case; 62 fmd_buf_destroy(hdl, NULL, page->page_bufname); 64 gmem_fmri_fini(hdl, &page->page_asru, destroy); 66 gmem_list_delete(&gmem.gm_pages, page); 67 fmd_hdl_free(hdl, page, sizeo 71 gmem_page_destroy(fmd_hdl_t *hdl, gmem_page_t *page) argument 80 gmem_page_t *page; local 95 gmem_page_t *page; local 158 gmem_page_t *page; local 174 gmem_page_t *page; local 248 gmem_page_unusable(fmd_hdl_t *hdl, gmem_page_t *page) argument 284 gmem_page_t *page, *next; local 295 gmem_page_dirty(fmd_hdl_t *hdl, gmem_page_t *page) argument 309 gmem_page_t *page; local 320 gmem_page_t *page = NULL; local [all...] |
/illumos-gate/usr/src/uts/common/io/ib/clients/rdsv3/ |
H A D | page.c | 6 * This file contains code imported from the OFED rds source file page.c 7 * Oracle elects to have and use the contents of page.c under and governed 58 caddr_t page; local 64 page = kmem_alloc(PAGE_SIZE, gfp); 65 if (!page) { 68 rdsv3_sg_set_page(scat, page, PAGE_SIZE, 0); 77 page = kmem_alloc(bytes, KM_NOSLEEP); 78 if (!page) { 83 rdsv3_sg_set_page(scat, page, bytes, 0);
|
/illumos-gate/usr/src/cmd/fm/modules/common/cpumem-retire/ |
H A D | cma_page.c | 30 * repeatedly attempt to retire a given page, but will not let us know when the 31 * page has been retired. We therefore have to poll to see if the retirement 37 * page. When forced to retry, we initially schedule a retry at a configurable 40 * for that page will occur with t equal to the maximum interval value. We 46 * value appropriate for the newest page. 62 cma_page_free(fmd_hdl_t *hdl, cma_page_t *page) argument 64 nvlist_free(page->pg_asru); 65 nvlist_free(page->pg_rsrc); 66 fmd_hdl_free(hdl, page, sizeof (cma_page_t)); 70 * Retire the specified ASRU, referring to a memory page b 89 cma_page_t *page; local 205 page_retry(fmd_hdl_t *hdl, cma_page_t *page) argument 258 cma_page_t *page = *pagep; local 297 cma_page_t *page; local [all...] |
/illumos-gate/usr/src/lib/libmapmalloc/common/ |
H A D | textmem.c | 44 * Each call to mmap() creates a page. The pages are linked in a list. 45 * Each page is divided in blocks. There is at least one block in a page. 64 struct page *page; /* Backwards reference to page */ member in struct:block 70 struct page { struct 71 size_t size; /* Total page size (incl. header) */ 72 struct page *next; 80 #define HDR_PAGE (sizeof (struct page) 110 struct page *page; local 243 defrag(struct page *page) argument [all...] |
/illumos-gate/usr/src/cmd/sgs/rtld/common/ |
H A D | malloc.c | 37 * Each call to mmap() creates a page. The pages are linked in a list. 38 * Each page is divided in blocks. There is at least one block in a page. 53 struct page *page; /* Backwards reference to page */ member in struct:block 59 struct page { struct 60 size_t size; /* Total page size (incl. header) */ 61 struct page *next; 69 #define HDR_PAGE (sizeof (struct page) 105 struct page *page; local 162 struct page *page; local 320 struct page *page; local [all...] |
/illumos-gate/usr/src/cmd/dtrace/test/tst/i86xpv/xdt/ |
H A D | tst.memenable.ksh | 55 xdt:mem::page-grant-map, 56 xdt:mem::page-grant-unmap, 57 xdt:mem::page-grant-transfer
|
/illumos-gate/usr/src/uts/common/io/drm/ |
H A D | drm_cache.c | 44 drm_clflush_page(caddr_t page) argument 48 if (page == NULL) 52 clflush_insn(page + i);
|
/illumos-gate/usr/src/uts/common/vm/ |
H A D | pvn.h | 59 struct page *pvn_read_kluster(struct vnode *vp, u_offset_t off, 63 struct page *pvn_write_kluster(struct vnode *vp, struct page *pp, 66 void pvn_read_done(struct page *plist, int flags); 67 void pvn_write_done(struct page *plist, int flags); 68 void pvn_io_done(struct page *plist); 70 int (*putapage)(vnode_t *, struct page *, u_offset_t *, 74 int pvn_getdirty(struct page *pp, int flags); 78 struct page *[], size_t, struct seg *, 81 uint_t *protp, struct page **p [all...] |
H A D | kpm.h | 48 struct page *kpe_page; /* back pointer to (start) page */
|
H A D | hat.h | 67 #include <vm/page.h> 147 * load/lock the given page struct 150 * load/lock the given array of page structs 153 * load/lock the given page frame number 204 void hat_memload(struct hat *, caddr_t, struct page *, uint_t, uint_t); 205 void hat_memload_array(struct hat *, caddr_t, size_t, struct page **, 207 void hat_memload_region(struct hat *, caddr_t, struct page *, uint_t, 209 void hat_memload_array_region(struct hat *, caddr_t, size_t, struct page **, 240 caddr_t hat_kpm_mapin(struct page *, struct kpme *); 241 void hat_kpm_mapout(struct page *, struc [all...] |
/illumos-gate/usr/src/cmd/fm/modules/sun4u/cpumem-diagnosis/ |
H A D | cmd_dp_page.c | 29 * Support routines for managing potential page and bank faults that have 32 * a page case is created with a special subtype of CMD_PTR_DP_PAGE_DEFER. An 36 * are seen while the page is deferred. This information is used to determine 37 * if the page should be faulted if the fault should be skipped because an 38 * intervening datapath fault has occurred. If a page is faulted when it is 42 * immediately retire the page where a UE is detected and the retire may or 92 fmd_hdl_abort(hdl, "too many mcids for deferred page"); 98 dp_page_defer_create(fmd_hdl_t *hdl, cmd_page_t *page, uint64_t afar) argument 105 dpage->dp_defer_page = page; 118 dp_page_defer_lookup(cmd_page_t *page) argument 136 cmd_page_t *page = cmd_page_lookup(afar); local 179 cmd_page_t *page; local 227 cmd_dp_page_restore(fmd_hdl_t *hdl, cmd_page_t *page) argument 244 cmd_page_t *page; local 262 cmd_dp_page_isdeferred(fmd_hdl_t *hdl, cmd_page_t *page) argument [all...] |
/illumos-gate/usr/src/cmd/lp/filter/postscript/dpost/ |
H A D | ps_include.c | 62 * page_no physical page number from *fin 76 /* found the page when non zero */ 82 /* prologue, page, and trailer offsets */ 83 Section prolog, page, trailer; local 101 * page. After the whole file is scanned, the special ps_include 103 * prologue, global definitions, the requested page, and the 120 page.start = page.end = 0; 129 page.start = ftell(fin); 133 else if (foundpage && page [all...] |
/illumos-gate/usr/src/uts/common/sys/ |
H A D | hold_page.h | 36 #include <vm/page.h> 42 * check the validity and make sure a page is not given away while it is mapped. 46 * If lock is PLAT_HOLD_NO_LOCK, simply check if the page pfn is valid 47 * in the system. If the page is valid, PLAT_HOLD_OK will be returned. 51 * the page exclusively. Again, if the lock is successful, the page 54 * page wasn't found, or the lock couldn't be grabbed, the return value 59 * Unlock the page pp. Should only be called after a previous,
|
/illumos-gate/usr/src/lib/libeti/form/common/ |
H A D | post.c | 114 P(f) = c->page; 126 if (c -> page != P(f)) { /* page change */ 128 v = _set_form_page(f, c->page, c); 158 set_form_page(FORM *f, int page) argument 160 if (!f || !ValidPage(f, page)) 164 P(f) = page; 171 if (page != P(f)) { 177 v = _set_form_page(f, page, (FIELD *) 0);
|
H A D | form.c | 62 (_PAGE *) 0, /* page */ 111 /* sort_form - sort fields on form(per page) */ 117 int i, page, pmin, pmax; local 121 for (page = 0; page < f->maxpage; ++page) { /* for each page */ 124 pmin = Pmin(f, page); 125 pmax = Pmax(f, page); 129 field[i]->page 175 _PAGE * page; local [all...] |
/illumos-gate/usr/src/cmd/more/ |
H A D | Makefile | 49 $(RM) $(ROOTBIN)/page 50 $(LN) $(ROOTPROG) $(ROOTBIN)/page
|
/illumos-gate/usr/src/uts/common/inet/nca/ |
H A D | ncakmem.h | 34 #include <vm/page.h>
|
/illumos-gate/usr/src/uts/sun4/os/ |
H A D | mem_config_arch.c | 32 #include <vm/page.h>
|
/illumos-gate/usr/src/uts/i86pc/os/ |
H A D | mem_config_arch.c | 33 #include <vm/page.h>
|
/illumos-gate/usr/src/uts/sun4u/starfire/os/ |
H A D | starfire.c | 40 #include <vm/page.h> 268 page_t *page, **freelist; local 276 /* for each page size */ 282 * page size. 314 page = *freelist; 315 result = page_trylock(page, SE_EXCL); 321 page_vpsub(freelist, page); 323 mach_page_sub(freelist, page); 327 lbn = PFN_2_LBN(page->p_pagenum); 331 page_vpadd(&bdlist[lbn], page); [all...] |
/illumos-gate/usr/src/cmd/sendmail/db/btree/ |
H A D | bt_split.c | 71 * Split a page. 95 * When __bam_split is first called, we know that a leaf page was too 96 * full for an insert. We don't know what leaf page it was, but we 98 * reacquire the leaf page, but this time get both the leaf page and 99 * its parent, locked. We then split the leaf page and see if the new 100 * internal key will fit into the parent page. If it will, we're done. 103 * only this time acquiring the parent page and its parent, locked. 105 * root page as the final resort. The entire process then repeats, 106 * as necessary, until we split a leaf page [all...] |