hxge_hw.c revision fe930412c257f961ae67039de3b164b83717976a
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License (the "License").
* You may not use this file except in compliance with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright 2008 Sun Microsystems, Inc. All rights reserved.
* Use is subject to license terms.
*/
#pragma ident "%Z%%M% %I% %E% SMI"
#include <hxge_impl.h>
lb_property_t lb_normal = {normal, "normal", hxge_lb_normal};
lb_property_t lb_mac10g = {internal, "mac10g", hxge_lb_mac10g};
uint32_t hxge_lb_dbg = 1;
extern uint32_t hxge_jumbo_mtu;
extern boolean_t hxge_jumbo_enable;
static void hxge_rtrace_ioctl(p_hxge_t, queue_t *, mblk_t *, struct iocblk *);
void
hxge_global_reset(p_hxge_t hxgep)
{
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "==> hxge_global_reset"));
(void) hxge_intr_hw_disable(hxgep);
if (hxgep->suspended)
(void) hxge_link_init(hxgep);
(void) hxge_vmac_init(hxgep);
(void) hxge_intr_hw_enable(hxgep);
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "<== hxge_global_reset"));
}
void
hxge_hw_id_init(p_hxge_t hxgep)
{
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "==> hxge_hw_id_init"));
/*
* Set up initial hardware parameters required such as mac mtu size.
*/
hxgep->vmac.is_jumbo = B_FALSE;
/* 1518 + 4 + 16 */
hxgep->vmac.maxframesize = STD_FRAME_SIZE + TX_PKT_HEADER_SIZE;
if (hxgep->param_arr[param_accept_jumbo].value || hxge_jumbo_enable) {
hxgep->vmac.maxframesize = (uint16_t)hxge_jumbo_mtu;
hxgep->vmac.is_jumbo = B_TRUE;
}
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "==> hxge_hw_id_init: maxframesize %d",
hxgep->vmac.maxframesize));
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "<== hxge_hw_id_init"));
}
void
hxge_hw_init_niu_common(p_hxge_t hxgep)
{
p_hxge_hw_list_t hw_p;
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "==> hxge_hw_init_niu_common"));
if ((hw_p = hxgep->hxge_hw_p) == NULL) {
return;
}
MUTEX_ENTER(&hw_p->hxge_cfg_lock);
if (hw_p->flags & COMMON_INIT_DONE) {
HXGE_DEBUG_MSG((hxgep, MOD_CTL, "hxge_hw_init_niu_common"
" already done for dip $%p exiting", hw_p->parent_devp));
MUTEX_EXIT(&hw_p->hxge_cfg_lock);
return;
}
hw_p->flags = COMMON_INIT_START;
HXGE_DEBUG_MSG((hxgep, MOD_CTL,
"hxge_hw_init_niu_common Started for device id %x",
hw_p->parent_devp));
(void) hxge_pfc_hw_reset(hxgep);
hw_p->flags = COMMON_INIT_DONE;
MUTEX_EXIT(&hw_p->hxge_cfg_lock);
HXGE_DEBUG_MSG((hxgep, MOD_CTL,
"hxge_hw_init_niu_common Done for device id %x",
hw_p->parent_devp));
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "<== hxge_hw_init_niu_common"));
}
uint_t
hxge_intr(caddr_t arg1, caddr_t arg2)
{
p_hxge_ldv_t ldvp = (p_hxge_ldv_t)arg1;
p_hxge_t hxgep = (p_hxge_t)arg2;
uint_t serviced = DDI_INTR_UNCLAIMED;
uint8_t ldv;
hpi_handle_t handle;
p_hxge_ldgv_t ldgvp;
p_hxge_ldg_t ldgp, t_ldgp;
p_hxge_ldv_t t_ldvp;
uint32_t vector0 = 0, vector1 = 0;
int i, j, nldvs, nintrs = 1;
hpi_status_t rs = HPI_SUCCESS;
/*
* DDI interface returns second arg as NULL
*/
if ((arg2 == NULL) || ((void *) ldvp->hxgep != arg2)) {
hxgep = ldvp->hxgep;
}
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr"));
if (!(hxgep->drv_state & STATE_HW_INITIALIZED)) {
HXGE_ERROR_MSG((hxgep, INT_CTL,
"<== hxge_intr: not initialized 0x%x", serviced));
return (serviced);
}
ldgvp = hxgep->ldgvp;
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr: ldgvp $%p", ldgvp));
if (ldvp == NULL && ldgvp) {
t_ldvp = ldvp = ldgvp->ldvp;
}
if (ldvp) {
ldgp = t_ldgp = ldvp->ldgp;
}
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr: "
"ldgvp $%p ldvp $%p ldgp $%p", ldgvp, ldvp, ldgp));
if (ldgvp == NULL || ldvp == NULL || ldgp == NULL) {
HXGE_ERROR_MSG((hxgep, INT_CTL, "==> hxge_intr: "
"ldgvp $%p ldvp $%p ldgp $%p", ldgvp, ldvp, ldgp));
HXGE_ERROR_MSG((hxgep, INT_CTL, "<== hxge_intr: not ready"));
return (DDI_INTR_UNCLAIMED);
}
/*
* This interrupt handler will have to go through
* all the logical devices to find out which
* logical device interrupts us and then call
* its handler to process the events.
*/
handle = HXGE_DEV_HPI_HANDLE(hxgep);
t_ldgp = ldgp;
t_ldvp = ldgp->ldvp;
nldvs = ldgp->nldvs;
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr: #ldvs %d #intrs %d",
nldvs, ldgvp->ldg_intrs));
serviced = DDI_INTR_CLAIMED;
for (i = 0; i < nintrs; i++, t_ldgp++) {
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr(%d): #ldvs %d "
" #intrs %d", i, nldvs, nintrs));
/* Get this group's flag bits. */
t_ldgp->interrupted = B_FALSE;
rs = hpi_ldsv_ldfs_get(handle, t_ldgp->ldg, &vector0, &vector1);
if (rs) {
continue;
}
if (!vector0 && !vector1) {
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr: "
"no interrupts on group %d", t_ldgp->ldg));
continue;
}
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr: "
"vector0 0x%llx vector1 0x%llx", vector0, vector1));
t_ldgp->interrupted = B_TRUE;
nldvs = t_ldgp->nldvs;
for (j = 0; j < nldvs; j++, t_ldvp++) {
/*
* Call device's handler if flag bits are on.
*/
ldv = t_ldvp->ldv;
if ((LDV_ON(ldv, vector0) | (LDV_ON(ldv, vector1)))) {
HXGE_DEBUG_MSG((hxgep, INT_CTL,
"==> hxge_intr: calling device %d"
" #ldvs %d #intrs %d", j, nldvs, nintrs));
(void) (t_ldvp->ldv_intr_handler)(
(caddr_t)t_ldvp, arg2);
}
}
}
t_ldgp = ldgp;
for (i = 0; i < nintrs; i++, t_ldgp++) {
/* rearm group interrupts */
if (t_ldgp->interrupted) {
HXGE_DEBUG_MSG((hxgep, INT_CTL,
"==> hxge_intr: arm group %d", t_ldgp->ldg));
(void) hpi_intr_ldg_mgmt_set(handle, t_ldgp->ldg,
t_ldgp->arm, t_ldgp->ldg_timer);
}
}
HXGE_DEBUG_MSG((hxgep, INT_CTL, "<== hxge_intr: serviced 0x%x",
serviced));
return (serviced);
}
hxge_status_t
hxge_peu_handle_sys_errors(p_hxge_t hxgep)
{
hpi_handle_t handle;
p_hxge_peu_sys_stats_t statsp;
peu_intr_stat_t stat;
handle = hxgep->hpi_handle;
statsp = (p_hxge_peu_sys_stats_t)&hxgep->statsp->peu_sys_stats;
HXGE_REG_RD64(handle, PEU_INTR_STAT, &stat.value);
/*
* The PCIE errors are unrecoverrable and cannot be cleared.
* The only thing we can do here is to mask them off to prevent
* continued interrupts.
*/
HXGE_REG_WR64(handle, PEU_INTR_MASK, 0xffffffff);
if (stat.bits.spc_acc_err) {
statsp->spc_acc_err++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: spc_acc_err"));
}
if (stat.bits.tdc_pioacc_err) {
statsp->tdc_pioacc_err++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: tdc_pioacc_err"));
}
if (stat.bits.rdc_pioacc_err) {
statsp->rdc_pioacc_err++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: rdc_pioacc_err"));
}
if (stat.bits.pfc_pioacc_err) {
statsp->pfc_pioacc_err++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: pfc_pioacc_err"));
}
if (stat.bits.vmac_pioacc_err) {
statsp->vmac_pioacc_err++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: vmac_pioacc_err"));
}
if (stat.bits.cpl_hdrq_parerr) {
statsp->cpl_hdrq_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: cpl_hdrq_parerr"));
}
if (stat.bits.cpl_dataq_parerr) {
statsp->cpl_dataq_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: cpl_dataq_parerr"));
}
if (stat.bits.retryram_xdlh_parerr) {
statsp->retryram_xdlh_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: retryram_xdlh_parerr"));
}
if (stat.bits.retrysotram_xdlh_parerr) {
statsp->retrysotram_xdlh_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: retrysotram_xdlh_parerr"));
}
if (stat.bits.p_hdrq_parerr) {
statsp->p_hdrq_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: p_hdrq_parerr"));
}
if (stat.bits.p_dataq_parerr) {
statsp->p_dataq_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: p_dataq_parerr"));
}
if (stat.bits.np_hdrq_parerr) {
statsp->np_hdrq_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: np_hdrq_parerr"));
}
if (stat.bits.np_dataq_parerr) {
statsp->np_dataq_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: np_dataq_parerr"));
}
if (stat.bits.eic_msix_parerr) {
statsp->eic_msix_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: eic_msix_parerr"));
}
if (stat.bits.hcr_parerr) {
statsp->hcr_parerr++;
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_peu_handle_sys_errors: hcr_parerr"));
}
return (HXGE_OK);
}
/*ARGSUSED*/
uint_t
hxge_syserr_intr(caddr_t arg1, caddr_t arg2)
{
p_hxge_ldv_t ldvp = (p_hxge_ldv_t)arg1;
p_hxge_t hxgep = (p_hxge_t)arg2;
p_hxge_ldg_t ldgp = NULL;
hpi_handle_t handle;
dev_err_stat_t estat;
uint_t serviced = DDI_INTR_UNCLAIMED;
if ((arg1 == NULL) && (arg2 == NULL)) {
return (serviced);
}
if ((arg2 == NULL) ||
((ldvp != NULL) && ((void *)ldvp->hxgep != arg2))) {
if (ldvp != NULL) {
hxgep = ldvp->hxgep;
}
}
HXGE_DEBUG_MSG((hxgep, SYSERR_CTL,
"==> hxge_syserr_intr: arg2 $%p arg1 $%p", hxgep, ldvp));
if (ldvp != NULL && ldvp->use_timer == B_FALSE) {
ldgp = ldvp->ldgp;
if (ldgp == NULL) {
HXGE_ERROR_MSG((hxgep, SYSERR_CTL,
"<== hxge_syserrintr(no logical group): "
"arg2 $%p arg1 $%p", hxgep, ldvp));
return (DDI_INTR_UNCLAIMED);
}
/*
* Get the logical device state if the function uses interrupt.
*/
}
/* This interrupt handler is for system error interrupts. */
handle = HXGE_DEV_HPI_HANDLE(hxgep);
estat.value = 0;
(void) hpi_fzc_sys_err_stat_get(handle, &estat);
HXGE_DEBUG_MSG((hxgep, SYSERR_CTL,
"==> hxge_syserr_intr: device error 0x%016llx", estat.value));
if (estat.bits.tdc_err0 || estat.bits.tdc_err1) {
/* TDMC */
(void) hxge_txdma_handle_sys_errors(hxgep);
} else if (estat.bits.rdc_err0 || estat.bits.rdc_err1) {
/* RDMC */
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_syserr_intr: device error - RDMC"));
(void) hxge_rxdma_handle_sys_errors(hxgep);
} else if (estat.bits.vnm_pio_err1 || estat.bits.peu_err1) {
/* PCI-E */
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_syserr_intr: device error - PCI-E"));
/* kstats are updated here */
(void) hxge_peu_handle_sys_errors(hxgep);
if (estat.bits.peu_err1)
HXGE_FM_REPORT_ERROR(hxgep, NULL,
HXGE_FM_EREPORT_PEU_ERR);
if (estat.bits.vnm_pio_err1)
HXGE_FM_REPORT_ERROR(hxgep, NULL,
HXGE_FM_EREPORT_PEU_VNM_PIO_ERR);
} else if (estat.value != 0) {
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"==> hxge_syserr_intr: device error - unknown"));
}
serviced = DDI_INTR_CLAIMED;
if ((ldgp != NULL) && (ldvp != NULL) &&
(ldgp->nldvs == 1) && !ldvp->use_timer) {
(void) hpi_intr_ldg_mgmt_set(handle, ldgp->ldg,
B_TRUE, ldgp->ldg_timer);
}
HXGE_DEBUG_MSG((hxgep, SYSERR_CTL, "<== hxge_syserr_intr"));
return (serviced);
}
void
hxge_intr_hw_enable(p_hxge_t hxgep)
{
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr_hw_enable"));
(void) hxge_intr_mask_mgmt_set(hxgep, B_TRUE);
HXGE_DEBUG_MSG((hxgep, INT_CTL, "<== hxge_intr_hw_enable"));
}
void
hxge_intr_hw_disable(p_hxge_t hxgep)
{
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_intr_hw_disable"));
(void) hxge_intr_mask_mgmt_set(hxgep, B_FALSE);
HXGE_DEBUG_MSG((hxgep, INT_CTL, "<== hxge_intr_hw_disable"));
}
void
hxge_rx_hw_blank(void *arg, time_t ticks, uint_t count)
{
p_hxge_t hxgep = (p_hxge_t)arg;
HXGE_DEBUG_MSG((hxgep, INT_CTL, "==> hxge_rx_hw_blank"));
/*
* Replace current ticks and counts for later
* processing by the receive packet interrupt routines.
*/
hxgep->intr_timeout = (uint16_t)ticks;
hxgep->intr_threshold = (uint16_t)count;
HXGE_DEBUG_MSG((hxgep, INT_CTL, "<== hxge_rx_hw_blank"));
}
void
hxge_hw_stop(p_hxge_t hxgep)
{
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "==> hxge_hw_stop"));
(void) hxge_tx_vmac_disable(hxgep);
(void) hxge_rx_vmac_disable(hxgep);
(void) hxge_txdma_hw_mode(hxgep, HXGE_DMA_STOP);
(void) hxge_rxdma_hw_mode(hxgep, HXGE_DMA_STOP);
HXGE_DEBUG_MSG((hxgep, DDI_CTL, "<== hxge_hw_stop"));
}
void
hxge_hw_ioctl(p_hxge_t hxgep, queue_t *wq, mblk_t *mp, struct iocblk *iocp)
{
int cmd;
HXGE_DEBUG_MSG((hxgep, IOC_CTL, "==> hxge_hw_ioctl"));
if (hxgep == NULL) {
miocnak(wq, mp, 0, EINVAL);
return;
}
iocp->ioc_error = 0;
cmd = iocp->ioc_cmd;
switch (cmd) {
default:
miocnak(wq, mp, 0, EINVAL);
return;
case HXGE_GET64:
hxge_get64(hxgep, mp->b_cont);
miocack(wq, mp, sizeof (uint32_t), 0);
break;
case HXGE_PUT64:
hxge_put64(hxgep, mp->b_cont);
miocack(wq, mp, 0, 0);
break;
case HXGE_PUT_TCAM:
hxge_put_tcam(hxgep, mp->b_cont);
miocack(wq, mp, 0, 0);
break;
case HXGE_GET_TCAM:
hxge_get_tcam(hxgep, mp->b_cont);
miocack(wq, mp, 0, 0);
break;
case HXGE_RTRACE:
hxge_rtrace_ioctl(hxgep, wq, mp, iocp);
break;
}
}
/*
* 10G is the only loopback mode for Hydra.
*/
void
hxge_loopback_ioctl(p_hxge_t hxgep, queue_t *wq, mblk_t *mp,
struct iocblk *iocp)
{
p_lb_property_t lb_props;
size_t size;
int i;
if (mp->b_cont == NULL) {
miocnak(wq, mp, 0, EINVAL);
}
switch (iocp->ioc_cmd) {
case LB_GET_MODE:
HXGE_DEBUG_MSG((hxgep, IOC_CTL, "HXGE_GET_LB_MODE command"));
if (hxgep != NULL) {
*(lb_info_sz_t *)mp->b_cont->b_rptr =
hxgep->statsp->port_stats.lb_mode;
miocack(wq, mp, sizeof (hxge_lb_t), 0);
} else
miocnak(wq, mp, 0, EINVAL);
break;
case LB_SET_MODE:
HXGE_DEBUG_MSG((hxgep, IOC_CTL, "HXGE_SET_LB_MODE command"));
if (iocp->ioc_count != sizeof (uint32_t)) {
miocack(wq, mp, 0, 0);
break;
}
if ((hxgep != NULL) && hxge_set_lb(hxgep, wq, mp->b_cont)) {
miocack(wq, mp, 0, 0);
} else {
miocnak(wq, mp, 0, EPROTO);
}
break;
case LB_GET_INFO_SIZE:
HXGE_DEBUG_MSG((hxgep, IOC_CTL, "LB_GET_INFO_SIZE command"));
if (hxgep != NULL) {
size = sizeof (lb_normal) + sizeof (lb_mac10g);
*(lb_info_sz_t *)mp->b_cont->b_rptr = size;
HXGE_DEBUG_MSG((hxgep, IOC_CTL,
"HXGE_GET_LB_INFO command: size %d", size));
miocack(wq, mp, sizeof (lb_info_sz_t), 0);
} else
miocnak(wq, mp, 0, EINVAL);
break;
case LB_GET_INFO:
HXGE_DEBUG_MSG((hxgep, IOC_CTL, "HXGE_GET_LB_INFO command"));
if (hxgep != NULL) {
size = sizeof (lb_normal) + sizeof (lb_mac10g);
HXGE_DEBUG_MSG((hxgep, IOC_CTL,
"HXGE_GET_LB_INFO command: size %d", size));
if (size == iocp->ioc_count) {
i = 0;
lb_props = (p_lb_property_t)mp->b_cont->b_rptr;
lb_props[i++] = lb_normal;
lb_props[i++] = lb_mac10g;
miocack(wq, mp, size, 0);
} else
miocnak(wq, mp, 0, EINVAL);
} else {
miocnak(wq, mp, 0, EINVAL);
cmn_err(CE_NOTE, "hxge_hw_ioctl: invalid command 0x%x",
iocp->ioc_cmd);
}
break;
}
}
/*ARGSUSED*/
boolean_t
hxge_set_lb(p_hxge_t hxgep, queue_t *wq, p_mblk_t mp)
{
boolean_t status = B_TRUE;
uint32_t lb_mode;
lb_property_t *lb_info;
HXGE_DEBUG_MSG((hxgep, IOC_CTL, "<== hxge_set_lb"));
lb_mode = hxgep->statsp->port_stats.lb_mode;
if (lb_mode == *(uint32_t *)mp->b_rptr) {
cmn_err(CE_NOTE,
"hxge%d: Loopback mode already set (lb_mode %d).\n",
hxgep->instance, lb_mode);
status = B_FALSE;
goto hxge_set_lb_exit;
}
lb_mode = *(uint32_t *)mp->b_rptr;
lb_info = NULL;
/* 10G is the only loopback mode for Hydra */
if (lb_mode == lb_normal.value)
lb_info = &lb_normal;
else if (lb_mode == lb_mac10g.value)
lb_info = &lb_mac10g;
else {
cmn_err(CE_NOTE,
"hxge%d: Loopback mode not supported(mode %d).\n",
hxgep->instance, lb_mode);
status = B_FALSE;
goto hxge_set_lb_exit;
}
if (lb_mode == hxge_lb_normal) {
if (hxge_lb_dbg) {
cmn_err(CE_NOTE,
"!hxge%d: Returning to normal operation",
hxgep->instance);
}
hxgep->statsp->port_stats.lb_mode = hxge_lb_normal;
hxge_global_reset(hxgep);
goto hxge_set_lb_exit;
}
hxgep->statsp->port_stats.lb_mode = lb_mode;
if (hxge_lb_dbg)
cmn_err(CE_NOTE, "!hxge%d: Adapter now in %s loopback mode",
hxgep->instance, lb_info->key);
if (lb_info->lb_type == internal) {
if ((hxgep->statsp->port_stats.lb_mode == hxge_lb_mac10g))
hxgep->statsp->mac_stats.link_speed = 10000;
else {
cmn_err(CE_NOTE,
"hxge%d: Loopback mode not supported(mode %d).\n",
hxgep->instance, lb_mode);
status = B_FALSE;
goto hxge_set_lb_exit;
}
hxgep->statsp->mac_stats.link_duplex = 2;
hxgep->statsp->mac_stats.link_up = 1;
}
hxge_global_reset(hxgep);
hxge_set_lb_exit:
HXGE_DEBUG_MSG((hxgep, DDI_CTL,
"<== hxge_set_lb status = 0x%08x", status));
return (status);
}
void
hxge_check_hw_state(p_hxge_t hxgep)
{
p_hxge_ldgv_t ldgvp;
p_hxge_ldv_t t_ldvp;
HXGE_DEBUG_MSG((hxgep, SYSERR_CTL, "==> hxge_check_hw_state"));
MUTEX_ENTER(hxgep->genlock);
hxgep->hxge_timerid = 0;
if (!(hxgep->drv_state & STATE_HW_INITIALIZED)) {
goto hxge_check_hw_state_exit;
}
hxge_check_tx_hang(hxgep);
ldgvp = hxgep->ldgvp;
if (ldgvp == NULL || (ldgvp->ldvp_syserr == NULL)) {
HXGE_ERROR_MSG((hxgep, SYSERR_CTL, "<== hxge_check_hw_state: "
"NULL ldgvp (interrupt not ready)."));
goto hxge_check_hw_state_exit;
}
t_ldvp = ldgvp->ldvp_syserr;
if (!t_ldvp->use_timer) {
HXGE_DEBUG_MSG((hxgep, SYSERR_CTL, "<== hxge_check_hw_state: "
"ldgvp $%p t_ldvp $%p use_timer flag %d",
ldgvp, t_ldvp, t_ldvp->use_timer));
goto hxge_check_hw_state_exit;
}
if (fm_check_acc_handle(hxgep->dev_regs->hxge_regh) != DDI_FM_OK) {
HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
"Bad register acc handle"));
}
(void) hxge_syserr_intr((caddr_t)t_ldvp, (caddr_t)hxgep);
hxgep->hxge_timerid = hxge_start_timer(hxgep, hxge_check_hw_state,
HXGE_CHECK_TIMER);
hxge_check_hw_state_exit:
MUTEX_EXIT(hxgep->genlock);
HXGE_DEBUG_MSG((hxgep, SYSERR_CTL, "<== hxge_check_hw_state"));
}
/*ARGSUSED*/
static void
hxge_rtrace_ioctl(p_hxge_t hxgep, queue_t *wq, mblk_t *mp,
struct iocblk *iocp)
{
ssize_t size;
rtrace_t *rtp;
mblk_t *nmp;
uint32_t i, j;
uint32_t start_blk;
uint32_t base_entry;
uint32_t num_entries;
HXGE_DEBUG_MSG((hxgep, STR_CTL, "==> hxge_rtrace_ioctl"));
size = 1024;
if (mp->b_cont == NULL || MBLKL(mp->b_cont) < size) {
HXGE_DEBUG_MSG((hxgep, STR_CTL,
"malformed M_IOCTL MBLKL = %d size = %d",
MBLKL(mp->b_cont), size));
miocnak(wq, mp, 0, EINVAL);
return;
}
nmp = mp->b_cont;
rtp = (rtrace_t *)nmp->b_rptr;
start_blk = rtp->next_idx;
num_entries = rtp->last_idx;
base_entry = start_blk * MAX_RTRACE_IOC_ENTRIES;
HXGE_DEBUG_MSG((hxgep, STR_CTL, "start_blk = %d\n", start_blk));
HXGE_DEBUG_MSG((hxgep, STR_CTL, "num_entries = %d\n", num_entries));
HXGE_DEBUG_MSG((hxgep, STR_CTL, "base_entry = %d\n", base_entry));
rtp->next_idx = hpi_rtracebuf.next_idx;
rtp->last_idx = hpi_rtracebuf.last_idx;
rtp->wrapped = hpi_rtracebuf.wrapped;
for (i = 0, j = base_entry; i < num_entries; i++, j++) {
rtp->buf[i].ctl_addr = hpi_rtracebuf.buf[j].ctl_addr;
rtp->buf[i].val_l32 = hpi_rtracebuf.buf[j].val_l32;
rtp->buf[i].val_h32 = hpi_rtracebuf.buf[j].val_h32;
}
nmp->b_wptr = nmp->b_rptr + size;
HXGE_DEBUG_MSG((hxgep, STR_CTL, "<== hxge_rtrace_ioctl"));
miocack(wq, mp, (int)size, 0);
}