Lines Matching defs:vecp

651 	apix_vector_t *vecp;
667 vecp = xv_vector(cpuid, vector);
668 if (vecp == NULL) {
674 nipl = vecp->v_pri;
692 ASSERT(vecp->v_state != APIX_STATE_OBSOLETED);
696 (vecp->v_type & APIX_TYPE_FIXED) && apic_level_intr[vecp->v_inum])
697 apix_level_intr_pre_eoi(vecp->v_inum);
713 APIC_DEBUG_BUF_PUT(vecp->v_inum);
782 apix_vector_t *vecp = xv_vector(cpuid, vector);
785 ASSERT(vecp != NULL && LOCK_HELD(&apix_lock));
787 if (vecp->v_type == APIX_TYPE_FIXED)
788 apix_intx_set_shared(vecp->v_inum, 1);
791 if (vecp->v_share > 1)
795 if (vecp->v_type == APIX_TYPE_IPI)
805 (void) apix_setup_io_intr(vecp);
815 apix_vector_t *vecp = xv_vector(cpuid, vector);
818 ASSERT(vecp != NULL && LOCK_HELD(&apix_lock));
820 if (vecp->v_type == APIX_TYPE_FIXED)
821 apix_intx_set_shared(vecp->v_inum, -1);
824 if (vecp->v_share > 1)
828 if (vecp->v_type == APIX_TYPE_IPI)
837 apix_disable_vector(vecp);
852 apix_vector_t *vecp, *newp;
863 vecp = apixp->x_vectbl[i];
864 if (!IS_VECT_ENABLED(vecp))
867 apix_disable_vector(vecp);
874 vecp = apixp->x_vectbl[i];
875 if (!IS_VECT_ENABLED(vecp))
878 if (vecp->v_flags & APIX_VECT_USER_BOUND) {
882 type = vecp->v_type;
893 newp = apix_set_cpu(vecp, bindcpu, &ret);
895 newp = apix_grp_set_cpu(vecp, bindcpu, &ret);
901 vecp->v_cpuid, vecp->v_vector);
923 apix_vector_t *vecp;
934 vecp = xv_vector(cpun, i);
935 if (!IS_VECT_ENABLED(vecp))
938 apix_enable_vector(vecp);
949 vecp = xv_vector(n, i);
950 if (!IS_VECT_ENABLED(vecp) ||
951 vecp->v_bound_cpuid != cpun)
954 if (vecp->v_type != APIX_TYPE_MSI)
955 (void) apix_set_cpu(vecp, cpun, &ret);
957 (void) apix_grp_set_cpu(vecp, cpun, &ret);
1142 apix_vector_t *vecp, *newvecp;
1189 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1190 if (vecp != NULL) {
1191 *result = APIX_VIRTVECTOR(vecp->v_cpuid,
1192 vecp->v_vector);
1203 vecp = apix_intx_xlate_vector(dip, hdlp->ih_inum,
1205 *result = (vecp == NULL) ? APIX_INVALID_VECT :
1206 APIX_VIRTVECTOR(vecp->v_cpuid, vecp->v_vector);
1212 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1213 if (vecp == NULL)
1216 *result = apix_get_pending(vecp);
1222 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1223 if (vecp == NULL)
1226 apix_intx_clear_mask(vecp->v_inum);
1232 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1233 if (vecp == NULL)
1236 apix_intx_set_mask(vecp->v_inum);
1242 vecp = apix_get_dev_map(dip, hdlp->ih_inum, hdlp->ih_type);
1243 if (vecp == NULL)
1246 *result = apix_intx_get_shared(vecp->v_inum);
1277 vecp = apix_get_req_vector(hdlp, hdlp->ih_flags);
1278 if (!IS_VECT_ENABLED(vecp)) {
1289 newvecp = apix_set_cpu(vecp, target, result);
1291 newvecp = apix_grp_set_cpu(vecp, target, result);
1371 apix_vector_t *vecp;
1378 if ((vecp = xv_vector(i, j)) != NULL)
1379 vecp->v_busy = 0;
1444 apix_get_pending(apix_vector_t *vecp)
1450 affinity_set(vecp->v_cpuid);
1452 index = vecp->v_vector / 32;
1453 bit = vecp->v_vector % 32;
1460 if (!pending && vecp->v_type == APIX_TYPE_FIXED)
1461 pending = apix_intx_get_pending(vecp->v_inum);
1469 apix_vector_t *vecp;
1483 vecp = xv_vector(cpuid, APIX_VIRTVEC_VECTOR(virt_vec));
1486 vecp = apix_get_dev_map(hdlp->ih_dip, hdlp->ih_inum,
1493 return (vecp);
1500 apix_vector_t *vecp;
1504 vecp = apix_get_req_vector(hdlp, intr_params_p->avgi_req_flags);
1505 if (IS_VECT_FREE(vecp)) {
1513 intr_params_p->avgi_cpu_id = vecp->v_cpuid;
1523 intr_params_p->avgi_vector = vecp->v_vector;
1528 intr_params_p->avgi_num_devs = vecp->v_share;
1536 for (i = 0, av_dev = vecp->v_autovect; av_dev;
1570 for (i = 0, av_dev = vecp->v_autovect; av_dev;
1589 apix_set_cpu(apix_vector_t *vecp, int new_cpu, int *result)
1604 if (vecp->v_type == APIX_TYPE_MSI) {
1605 if (i_ddi_intr_get_current_nintrs(APIX_GET_DIP(vecp)) > 1)
1608 return (apix_grp_set_cpu(vecp, new_cpu, result));
1614 if (vecp->v_type == APIX_TYPE_MSIX && IS_VECT_ENABLED(vecp)) {
1615 if ((dip = APIX_GET_DIP(vecp)) == NULL)
1617 inum = vecp->v_devp->dv_inum;
1637 if ((newp = apix_rebind(vecp, new_cpu, 1)) == NULL)
1651 apix_grp_set_cpu(apix_vector_t *vecp, int new_cpu, int *result)
1654 uint32_t orig_cpu = vecp->v_cpuid;
1655 int orig_vect = vecp->v_vector;
1663 " newcpu:%x\n", vecp->v_cpuid, vecp->v_vector, new_cpu));
1669 if (vecp->v_type != APIX_TYPE_MSI) {
1674 if ((dip = APIX_GET_DIP(vecp)) == NULL)
1685 if (vecp->v_inum != apix_get_min_dev_inum(dip, vecp->v_type))
1700 (vp->v_cpuid != vecp->v_cpuid))) {
1730 if ((newp = apix_rebind(vecp, new_cpu, num_vectors)) != NULL)
1787 apix_vector_t *vecp = xv_vector(irqp->airq_cpu, irqp->airq_vector);
1798 apic_vt_ops->apic_intrmap_alloc_entry(&vecp->v_intrmap_private, NULL,
1799 vecp->v_type, 1, ioapicindex);
1800 apic_vt_ops->apic_intrmap_map_entry(vecp->v_intrmap_private,
1801 (void *)&irdt, vecp->v_type, 1);
1802 apic_vt_ops->apic_intrmap_record_rdt(vecp->v_intrmap_private, &irdt);
1811 vecp->v_state = APIX_STATE_ENABLED;
2514 apix_vector_t *vecp;
2519 if ((vecp = apix_alloc_intx(dip, inum, irqno)) == NULL)
2525 vecp->v_cpuid, vecp->v_vector));
2539 apix_vector_t *vecp;
2546 vecp = apix_intx_get_vector(irqno);
2548 return (vecp);