Lines Matching refs:mp

69 static void		dr_init_mem_unit_data(dr_mem_unit_t *mp);
85 dr_get_memlist(dr_mem_unit_t *mp)
91 PR_MEM("%s for %s...\n", f, mp->sbm_cm.sbdev_path);
101 if (mp->sbm_mlist) {
104 mlist = memlist_dup(mp->sbm_mlist);
106 uint64_t basepa = _ptob64(mp->sbm_basepfn);
111 basepa &= ~mp->sbm_alignment_mask;
126 toppa = mp->sbm_slice_top;
142 mp->sbm_cm.sbdev_id, &mlist);
144 DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
150 PR_MEM("%s: memlist for %s\n", f, mp->sbm_cm.sbdev_path);
165 dr_mem_unit_t *mp = (dr_mem_unit_t *)cp;
181 ml = dr_get_memlist(mp);
216 DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
218 dr_dev_err(CE_WARN, &mp->sbm_cm, rv);
223 mp->sbm_cm.sbdev_id, mc->ml_address, mc->ml_size);
225 DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
231 dr_init_mem_unit_data(mp);
234 if (mp->sbm_cm.sbdev_error != NULL) {
253 * the results in *mp.
256 dr_del_mlist_query(struct memlist *mlist, memquery_t *mp)
263 mp->phys_pages = 0;
264 mp->managed = 0;
265 mp->nonrelocatable = 0;
266 mp->first_nonrelocatable = 0;
267 mp->last_nonrelocatable = 0;
283 dr_mem_unit_t *mp;
299 mp = dr_get_mem_unit(bp, m);
301 if (mp->sbm_cm.sbdev_state == DR_STATE_EMPTY) {
306 if (mp->sbm_cm.sbdev_id == (drmachid_t)0)
310 err = drmach_status(mp->sbm_cm.sbdev_id, &pstat);
312 DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
321 msp->ms_cm.c_id.c_type = mp->sbm_cm.sbdev_type;
322 msp->ms_cm.c_id.c_unit = mp->sbm_cm.sbdev_unum;
323 msp->ms_cm.c_cond = mp->sbm_cm.sbdev_cond;
324 msp->ms_cm.c_busy = mp->sbm_cm.sbdev_busy | pstat.busy;
325 msp->ms_cm.c_time = mp->sbm_cm.sbdev_time;
326 msp->ms_cm.c_ostate = mp->sbm_cm.sbdev_ostate;
328 msp->ms_totpages = mp->sbm_npages;
329 msp->ms_basepfn = mp->sbm_basepfn;
330 msp->ms_pageslost = mp->sbm_pageslost;
333 if (mp->sbm_flags & DR_MFLAG_RESERVED)
334 p_mp = mp->sbm_peer;
367 if (mp->sbm_cm.sbdev_state != DR_STATE_UNCONFIGURED) {
370 ml = dr_get_memlist(mp);
384 mp->sbm_cm.sbdev_id)) {
397 if ((mp->sbm_flags & DR_MFLAG_SOURCE) &&
398 (mp->sbm_cm.sbdev_state == DR_STATE_UNREFERENCED ||
399 mp->sbm_cm.sbdev_state == DR_STATE_RELEASE))
421 dr_mem_unit_t *mp = (dr_mem_unit_t *)devlist[d];
424 cmn_err(CE_CONT, "OS configure %s", mp->sbm_cm.sbdev_path);
426 state = mp->sbm_cm.sbdev_state;
430 f, mp->sbm_cm.sbdev_path);
433 ASSERT(mp->sbm_mlist != NULL);
435 f, mp->sbm_cm.sbdev_path);
436 PR_MEMLIST_DUMP(mp->sbm_mlist);
439 ASSERT((mp->sbm_flags & DR_MFLAG_RELOWNER) == 0);
445 f, mp->sbm_cm.sbdev_bp->b_path);
448 f, mp->sbm_cm.sbdev_path);
450 err = drmach_mem_enable(mp->sbm_cm.sbdev_id);
452 DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
458 dr_dev_err(CE_WARN, &mp->sbm_cm, ESBD_STATE);
481 dr_mem_unit_t *mp = (dr_mem_unit_t *)devlist[d];
484 mlist = dr_get_memlist(mp);
494 DR_DEV_INTERNAL_ERROR(&mp->sbm_cm);
497 f, mp->sbm_cm.sbdev_path);
508 mp->sbm_cm.sbdev_id,
512 DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
530 if (mp->sbm_mlist) {
531 memlist_delete(mp->sbm_mlist);
532 mp->sbm_mlist = NULL;
576 dr_disconnect_mem(dr_mem_unit_t *mp)
589 dr_init_mem_unit(dr_mem_unit_t *mp)
593 if (DR_DEV_IS_ATTACHED(&mp->sbm_cm)) {
595 mp->sbm_cm.sbdev_cond = SBD_COND_OK;
596 } else if (DR_DEV_IS_PRESENT(&mp->sbm_cm)) {
598 mp->sbm_cm.sbdev_cond = SBD_COND_OK;
599 } else if (mp->sbm_cm.sbdev_id != (drmachid_t)0) {
605 if (DR_DEV_IS_PRESENT(&mp->sbm_cm))
606 dr_init_mem_unit_data(mp);
609 dr_device_transition(&mp->sbm_cm, new_state);
613 dr_init_mem_unit_data(dr_mem_unit_t *mp)
615 drmachid_t id = mp->sbm_cm.sbdev_id;
623 ASSERT(mp->sbm_peer == NULL);
624 ASSERT(mp->sbm_flags == 0);
627 DRERR_SET_C(&mp->sbm_cm.sbdev_error, &err);
630 mp->sbm_basepfn = _b64top(minfo.mi_basepa);
631 mp->sbm_npages = _b64top(minfo.mi_size);
632 mp->sbm_alignment_mask = minfo.mi_alignment_mask;
633 mp->sbm_slice_base = minfo.mi_slice_base;
634 mp->sbm_slice_top = minfo.mi_slice_top;
635 mp->sbm_slice_size = minfo.mi_slice_size;
638 f, mp->sbm_cm.sbdev_path, mp->sbm_basepfn, mp->sbm_npages);