Searched defs:cqe (Results 1 - 25 of 27) sorted by relevance

12

/illumos-gate/usr/src/uts/common/io/fibre-channel/fca/oce/
H A Doce_mq.c42 struct oce_mq_cqe *cqe = NULL; local
50 /* do while we do not reach a cqe that is not valid */
55 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_mq_cqe);
56 while (cqe->u0.dw[3]) {
57 DW_SWAP(u32ptr(cqe), sizeof (struct oce_mq_cqe));
58 if (cqe->u0.s.async_event) {
59 acqe = (struct oce_async_cqe_link_state *)cqe;
76 cqe->u0.dw[3] = 0;
78 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_mq_cqe);
100 struct oce_mq_cqe *cqe local
[all...]
H A Doce_rx.c37 struct oce_nic_rx_cqe *cqe);
39 struct oce_rq *rq, struct oce_nic_rx_cqe *cqe);
42 static void oce_set_rx_oflags(mblk_t *mp, struct oce_nic_rx_cqe *cqe);
44 struct oce_nic_rx_cqe *cqe);
328 * cqe - Pointer to Completion Q entry
333 oce_rx(struct oce_dev *dev, struct oce_rq *rq, struct oce_nic_rx_cqe *cqe) argument
346 frag_cnt = cqe->u0.s.num_fragments & 0x7;
354 pkt_len = cqe->u0.s.pkt_size;
393 oce_rx_bcopy(struct oce_dev *dev, struct oce_rq *rq, struct oce_nic_rx_cqe *cqe) argument
410 pkt_len = cqe
440 oce_set_rx_oflags(mblk_t *mp, struct oce_nic_rx_cqe *cqe) argument
472 oce_rx_drop_pkt(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe) argument
496 struct oce_nic_rx_cqe *cqe; local
617 struct oce_nic_rx_cqe *cqe; local
[all...]
H A Doce_tx.c558 struct oce_nic_tx_cqe *cqe; local
571 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring, struct oce_nic_tx_cqe);
572 while (WQ_CQE_VALID(cqe)) {
574 DW_SWAP(u32ptr(cqe), sizeof (struct oce_nic_tx_cqe));
577 if (cqe->u0.s.status != 0) {
588 /* clear the valid bit and progress cqe */
589 WQ_CQE_INVALIDATE(cqe);
591 cqe = RING_GET_CONSUMER_ITEM_VA(cq->ring,
619 /* do while we do not reach a cqe that is not valid */
/illumos-gate/usr/src/uts/common/io/bnxe/577xx/drivers/common/lm/l4/
H A Dlm_l4fp.c263 * returns the next cqe in the cqe_buffer and updates the buffer params
267 char * cqe; local
269 cqe = cqe_buffer->head;
271 if(cqe == cqe_buffer->last) {
274 cqe_buffer->head = cqe + cqe_buffer->qe_size;
280 return cqe;
284 * returns the next occupied cqe in the cqe_buffer and updates the buffer params
289 char * cqe; local
291 cqe = cqe_buffer->tail;
293 if ((cqe
316 char * cqe; local
[all...]
H A Dlm_l4tx.c371 struct toe_tx_cqe * cqe,
377 /* get the cmd from cqe */
378 cmd = ((cqe->params & TOE_TX_CQE_COMPLETION_OPCODE) >> TOE_TX_CQE_COMPLETION_OPCODE_SHIFT);
382 /* Check that the cqe len make sense, we could have got here by chance... */
383 DbgBreakIfAll(cqe->len & 0xc0000000); /* two upper bits on show a completion larger than 1GB - a bit odd...*/
386 * fast-path part (nbytes completed) which will be handled in any case that cqe->len > 0 */
389 if (cqe->len &&
394 on RST recv cqe and do so only later on one of the following ramrod completions,
395 we need to ignore this too late completed bytes thus we nullify cqe->len */
400 lm_tcp_tx_inc_trm_aborted_bytes(pdev, tcp, cqe
369 lm_tcp_tx_process_cqe( lm_device_t * pdev, struct toe_tx_cqe * cqe, lm_tcp_state_t * tcp ) argument
455 struct toe_tx_cqe *cqe, *hist_cqe; local
[all...]
H A Dlm_l4rx.c1179 * processes a single cqe.
1183 struct toe_rx_cqe * cqe,
1191 cmd = ((cqe->params1 & TOE_RX_CQE_COMPLETION_OPCODE) >> TOE_RX_CQE_COMPLETION_OPCODE_SHIFT);
1194 /* Check that the cqe nbytes make sense, we could have got here by chance... */
1197 nbytes = (cqe->data.ooo_params.ooo_params & TOE_RX_CQE_OOO_PARAMS_NBYTES) >> TOE_RX_CQE_OOO_PARAMS_NBYTES_SHIFT;
1198 isle_num = (cqe->data.ooo_params.ooo_params & TOE_RX_CQE_OOO_PARAMS_ISLE_NUM) >> TOE_RX_CQE_OOO_PARAMS_ISLE_NUM_SHIFT;
1204 nbytes = cqe->data.raw_data;
1206 nbytes = (cqe->data.in_order_params.in_order_params & TOE_RX_CQE_IN_ORDER_PARAMS_NBYTES) >> TOE_RX_CQE_IN_ORDER_PARAMS_NBYTES_SHIFT;
1221 //DbgMessage(pdev, WARN, "GenericAdd cid=%d nbytes=%d!\n", tcp->cid, cqe->nbytes);
1254 //DbgMessage(pdev, WARN, "GenericRelease cid=%d nbytes=%d!\n", tcp->cid, cqe
1181 lm_tcp_rx_process_cqe( lm_device_t * pdev, struct toe_rx_cqe * cqe, lm_tcp_state_t * tcp, u8_t sb_idx) argument
1427 struct toe_rx_cqe *cqe, *hist_cqe; local
[all...]
/illumos-gate/usr/src/uts/common/io/ib/mgt/ibmf/
H A Dibmf_handlers.c313 ibt_wc_t cqe; local
330 status = ibt_poll_cq(cq_handle, &cqe, 1, NULL);
348 ibmf_i_process_completion(ibmf_cip, &cqe);
358 status = ibt_poll_cq(cq_handle, &cqe, 1, NULL);
376 ibmf_i_process_completion(ibmf_cip, &cqe);
/illumos-gate/usr/src/lib/udapl/udapl_tavor/common/
H A Ddapl_evd_util.c55 IN ib_work_completion_t cqe);
405 * cqe
417 IN ib_work_completion_t cqe)
433 dto_cookie = (DAPL_COOKIE *) (uintptr_t)DAPL_GET_CQE_WRID(&cqe);
440 "\t\t work_req_id 0x%llx\n", DAPL_GET_CQE_WRID(&cqe));
442 "\t\t op_type: %s\n", optable[DAPL_GET_CQE_OPTYPE(&cqe)]);
443 if ((DAPL_GET_CQE_OPTYPE(&cqe) == OP_SEND) ||
444 (DAPL_GET_CQE_OPTYPE(&cqe) == OP_RDMA_WRITE)) {
449 "\t\t bytes_num %d\n", DAPL_GET_CQE_BYTESNUM(&cqe));
452 "\t\t status %d\n", DAPL_GET_CQE_STATUS(&cqe));
416 dapli_evd_eh_print_cqe( IN ib_work_completion_t cqe) argument
844 ib_work_completion_t *cqe; local
1220 ib_work_completion_t cqe[MAX_CQES_PER_POLL]; local
[all...]
/illumos-gate/usr/src/uts/common/io/bnxe/577xx/drivers/common/lm/device/
H A Dlm_recv.c338 * @param cqe
350 IN const struct eth_end_agg_rx_cqe* cqe,
359 u32_t sge_size = mm_le16_to_cpu(cqe->pkt_len) - pkt->l2pkt_rx_info->size;
372 DbgBreakIf( mm_le16_to_cpu(cqe->pkt_len) < pkt->l2pkt_rx_info->size);
380 pkt->l2pkt_rx_info->total_packet_size = mm_le16_to_cpu(cqe->pkt_len);
381 pkt->l2pkt_rx_info->coal_seg_cnt = mm_le16_to_cpu(cqe->num_of_coalesced_segs);
382 pkt->l2pkt_rx_info->dup_ack_cnt = cqe->pure_ack_count;
383 pkt->l2pkt_rx_info->ts_delta = mm_le32_to_cpu(cqe->timestamp_delta);
392 ASSERT_STATIC(LM_TPA_MAX_AGG_SIZE == ARRSIZE(cqe->sgl_or_raw_data.sgl));
393 DbgBreakIf(ARRSIZE(cqe
348 lm_tpa_stop( IN lm_device_t* pdev, INOUT s_list_t* rcvd_list, IN const struct eth_end_agg_rx_cqe* cqe, IN const u32_t chain_idx, IN u32_t pkt_cnt, IN const u8_t queue_index) argument
501 lm_tpa_start_flags_handle( IN lm_device_t* pdev, IN const struct eth_fast_path_rx_cqe* cqe, INOUT lm_packet_t* pkt, IN const u16_t parse_flags) argument
569 lm_regular_flags_handle( IN lm_device_t* pdev, IN const struct eth_fast_path_rx_cqe* cqe, INOUT lm_packet_t* pkt, IN const u16_t parse_flags) argument
933 union eth_rx_cqe* cqe = NULL; local
[all...]
H A Dlm_sp.c3362 /* get hold of the cqe, and find out what it's type corresponds to */
3382 /* Recycle the cqe */
4046 void lm_eth_init_command_comp(struct _lm_device_t *pdev, struct common_ramrod_eth_rx_cqe *cqe) argument
4050 u32_t conn_and_cmd_data = mm_le32_to_cpu(cqe->conn_and_cmd_data);
4053 u8_t ramrod_type = cqe->ramrod_type;
4099 empty_data = mm_le32_to_cpu(cqe->protocol_data.data_lo);
4163 mm_eth_ramrod_comp_cb(pdev, cqe);
4178 struct common_ramrod_eth_rx_cqe cqe; local
4180 /* The idea is to prepare a cqe and call: common_ramrod_eth_rx_cqe */
4181 cqe
[all...]
/illumos-gate/usr/src/lib/udapl/udapl_tavor/tavor/
H A Ddapl_tavor_wr.c59 dapls_tavor_wrid_get_entry(ib_cq_handle_t cq, tavor_hw_cqe_t *cqe, argument
71 qpnum = TAVOR_CQE_QPNUM_GET(cqe);
87 wre_tmp = dapli_tavor_wrid_find_match(wq, cqe);
124 dapli_tavor_wrid_find_match(dapls_tavor_workq_hdr_t *wq, tavor_hw_cqe_t *cqe) argument
135 wqeaddr_size = TAVOR_CQE_WQEADDRSZ_GET(cqe);
161 curr = dapli_tavor_wrid_find_match_srq(container, cqe);
257 tavor_hw_cqe_t *cqe)
266 wqe_addr = TAVOR_CQE_WQEADDRSZ_GET(cqe) & 0xFFFFFFC0;
256 dapli_tavor_wrid_find_match_srq(dapls_tavor_wrid_list_hdr_t *wl, tavor_hw_cqe_t *cqe) argument
H A Ddapl_arbel_hw.c554 tavor_hw_cqe_t *cqe; local
572 cqe = &cq->cq_addr[cons_indx];
579 while (TAVOR_CQE_OWNER_IS_SW(cqe)) {
580 opcode = TAVOR_CQE_OPCODE_GET(cqe);
585 TAVOR_CQE_IMM_ETH_PKEY_CRED_GET(cqe);
596 cqe = &cq->cq_addr[cons_indx];
614 tavor_hw_cqe_t *cqe; local
631 cqe = &cq->cq_addr[cons_indx];
644 while (TAVOR_CQE_OWNER_IS_SW(cqe)) {
645 status = dapli_arbel_cq_cqe_consume(cq, cqe,
706 tavor_hw_cqe_t *cqe; local
752 dapli_arbel_cq_cqe_consume(ib_cq_handle_t cqhdl, tavor_hw_cqe_t *cqe, ibt_wc_t *wc) argument
861 dapli_arbel_cq_errcqe_consume(ib_cq_handle_t cqhdl, tavor_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1388 tavor_hw_cqe_t *cqe; local
[all...]
H A Ddapl_hermon_hw.c42 #define HERMON_CQE_OPCODE_GET(cqe) (((uint8_t *)cqe)[31] & 0x1F)
43 #define HERMON_CQE_SENDRECV_GET(cqe) (((uint8_t *)cqe)[31] & 0x40)
44 #define HERMON_CQE_OWNER_IS_SW(cq, cqe) ((((uint8_t *)cqe)[31] >> 7) == \
531 uint32_t *cqe; local
542 cqe = (uint32_t *)&cq->cq_addr[cons_indx];
549 while (HERMON_CQE_OWNER_IS_SW(cq, cqe)) {
550 opcode = HERMON_CQE_OPCODE_GET(cqe);
618 uint32_t *cqe; local
709 uint32_t *cqe; local
759 dapli_hermon_cq_cqe_consume(ib_cq_handle_t cqhdl, uint32_t *cqe, ibt_wc_t *wc) argument
862 dapli_hermon_cq_errcqe_consume(ib_cq_handle_t cqhdl, uint32_t *cqe, ibt_wc_t *wc) argument
1443 tavor_hw_cqe_t *cqe; local
[all...]
H A Ddapl_tavor_hw.c786 tavor_hw_cqe_t *cqe; local
804 cqe = &cq->cq_addr[cons_indx];
811 while (TAVOR_CQE_OWNER_IS_SW(cqe)) {
812 opcode = TAVOR_CQE_OPCODE_GET(cqe);
817 TAVOR_CQE_IMM_ETH_PKEY_CRED_GET(cqe);
828 cqe = &cq->cq_addr[cons_indx];
843 tavor_hw_cqe_t *cqe; local
861 cqe = &cq->cq_addr[cons_indx];
874 while (TAVOR_CQE_OWNER_IS_SW(cqe)) {
875 status = dapli_tavor_cq_cqe_consume(cq, cqe,
943 tavor_hw_cqe_t *cqe; local
991 dapli_tavor_cq_cqe_consume(ib_cq_handle_t cqhdl, tavor_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1096 dapli_tavor_cq_errcqe_consume(ib_cq_handle_t cqhdl, tavor_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1721 tavor_hw_cqe_t *cqe; local
[all...]
/illumos-gate/usr/src/uts/common/io/ib/adapters/tavor/
H A Dtavor_cq.c50 tavor_hw_cqe_t *cqe, ibt_wc_t *wc);
52 tavor_hw_cqe_t *cqe, ibt_wc_t *wc);
53 static void tavor_cqe_sync(tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe,
910 tavor_hw_cqe_t *cqe; local
941 cqe = &cq->cq_buf[cons_indx];
944 tavor_cqe_sync(cq, cqe, DDI_DMA_SYNC_FORCPU);
957 while (TAVOR_CQE_OWNER_IS_SW(cq, cqe)) {
958 status = tavor_cq_cqe_consume(state, cq, cqe,
962 TAVOR_CQE_OWNER_SET_HW(cq, cqe);
965 tavor_cqe_sync(cq, cqe, DDI_DMA_SYNC_FORDE
1300 tavor_cq_cqe_consume(tavor_state_t *state, tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1472 tavor_cq_errcqe_consume(tavor_state_t *state, tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1636 tavor_cqe_sync(tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe, uint_t flag) argument
1736 tavor_hw_cqe_t *cqe; local
[all...]
H A Dtavor_wr.c74 tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe);
2779 tavor_wrid_get_entry(tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe, argument
2796 opcode = TAVOR_CQE_OPCODE_GET(cq, cqe);
2804 send_or_recv = TAVOR_CQE_SENDRECV_GET(cq, cqe);
2808 qpnum = TAVOR_CQE_QPNUM_GET(cq, cqe);
2823 wre_tmp = tavor_wrid_find_match(wq, cq, cqe);
2861 tavor_hw_cqe_t *cqe)
2874 wqeaddr_size = TAVOR_CQE_WQEADDRSZ_GET(cq, cqe);
2899 curr = tavor_wrid_find_match_srq(container, cq, cqe);
2999 tavor_hw_cqe_t *cqe)
2860 tavor_wrid_find_match(tavor_workq_hdr_t *wq, tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe) argument
2998 tavor_wrid_find_match_srq(tavor_wrid_list_hdr_t *wl, tavor_cqhdl_t cq, tavor_hw_cqe_t *cqe) argument
[all...]
/illumos-gate/usr/src/uts/common/io/ib/clients/of/sol_ofs/
H A Dsol_kverbs.c929 * @cqe: The minimum size of the CQ.
940 int cqe, void *comp_vector)
951 "event_handler: 0x%p, cq_context: 0x%p, cqe: 0x%x, "
953 comp_handler, event_handler, cq_context, cqe, comp_vector);
962 "event_handler: 0x%p, cq_context: 0x%p, cqe: 0x%x, "
964 comp_handler, event_handler, cq_context, cqe, comp_vector,
971 "event_handler: 0x%p, cq_context: 0x%p, cqe: 0x%x, "
973 cq_context, cqe, comp_vector);
975 cq_attr.cq_size = cqe;
986 cq->cqe
938 ib_create_cq(struct ib_device *device, ib_comp_handler comp_handler, void (*event_handler)(struct ib_event *, void *), void *cq_context, int cqe, void *comp_vector) argument
[all...]
/illumos-gate/usr/src/uts/common/io/ib/adapters/hermon/
H A Dhermon_cq.c58 hermon_hw_cqe_t *cqe, ibt_wc_t *wc);
60 hermon_hw_cqe_t *cqe, ibt_wc_t *wc);
749 * in the cq_hdl, and setting up for the next cqe polling
870 hermon_hw_cqe_t *cqe; local
899 cqe = &cq->cq_buf[cons_indx & wrap_around_mask];
912 while (HERMON_CQE_OWNER_IS_SW(cq, cqe, cons_indx, shift, mask)) {
915 opcode = HERMON_CQE_OPCODE_GET(cq, cqe);
926 cqe = &cq->cq_buf[cons_indx & wrap_around_mask];
936 hermon_cq_cqe_consume(state, cq, cqe, &wc_p[polled_cnt++]);
942 cqe
1291 hermon_cq_cqe_consume(hermon_state_t *state, hermon_cqhdl_t cq, hermon_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1476 hermon_cq_errcqe_consume(hermon_state_t *state, hermon_cqhdl_t cq, hermon_hw_cqe_t *cqe, ibt_wc_t *wc) argument
1650 hermon_hw_cqe_t *cqe, *next_cqe; local
[all...]
H A Dhermon_wr.c2359 hermon_wrid_get_entry(hermon_cqhdl_t cq, hermon_hw_cqe_t *cqe) argument
2370 send_or_recv = HERMON_CQE_SENDRECV_GET(cq, cqe);
2373 qpnum = HERMON_CQE_QPNUM_GET(cq, cqe);
2387 indx = HERMON_CQE_WQEADDRSZ_GET(cq, cqe) & wq->wq_mask;
/illumos-gate/usr/src/uts/sun/sys/
H A Dsocal_cq_defs.h52 typedef struct cqe { struct
/illumos-gate/usr/src/uts/common/io/nvme/
H A Dnvme.c792 nvme_cqe_t *cqe; local
798 cqe = &qp->nq_cq[qp->nq_cqhead];
801 if (cqe->cqe_sf.sf_p == qp->nq_phase)
804 ASSERT(nvme->n_ioq[cqe->cqe_sqid] == qp);
805 ASSERT(cqe->cqe_cid < qp->nq_nentry);
808 cmd = qp->nq_cmd[cqe->cqe_cid];
809 qp->nq_cmd[cqe->cqe_cid] = NULL;
815 ASSERT(cmd->nc_sqid == cqe->cqe_sqid);
816 ASSERT(cmd->nc_sqe.sqe_cid == cqe->cqe_cid);
817 bcopy(cqe,
835 nvme_cqe_t *cqe = &cmd->nc_cqe; local
856 nvme_cqe_t *cqe = &cmd->nc_cqe; local
874 nvme_cqe_t *cqe = &cmd->nc_cqe; local
897 nvme_cqe_t *cqe = &cmd->nc_cqe; local
997 nvme_cqe_t *cqe = &cmd->nc_cqe; local
1096 nvme_cqe_t *cqe = &cmd->nc_cqe; local
[all...]
/illumos-gate/usr/src/uts/common/sys/ib/clients/of/rdma/
H A Dib_user_verbs.h310 uint32_t cqe; member in struct:ib_uverbs_create_cq
326 uint32_t cqe; member in struct:ib_uverbs_create_cq_resp
333 uint32_t cqe; member in struct:ib_uverbs_resize_cq
338 uint32_t cqe; member in struct:ib_uverbs_resize_cq_resp
H A Dib_verbs.h593 int cqe; member in struct:ib_cq
723 * @cqe: The minimum size of the CQ.
731 void *cq_context, int cqe, void *comp_vector);
/illumos-gate/usr/src/uts/sun/io/
H A Dsocal.c2614 socal_cq_enque(socal_state_t *socalp, socal_port_t *port_statep, cqe_t *cqe, argument
2632 p = (longlong_t *)cqe;
2719 cqe->cqe_hdr.cq_hdr_seqno = kcq->skc_seqno;
3296 cqe_t *cqe; local
3324 cqe = &(kcq->skc_cq[kcqv->skc_out]);
3345 srp = (soc_response_t *)cqe;
3557 cqe = &(kcq->skc_cq[kcqv->skc_out]);
3611 volatile cqe_t *cqe; local
3636 cqe = (volatile cqe_t *)&(kcq->skc_cq[kcqv->skc_out]);
3647 if ((hdr_count = cqe
3996 socal_us_els(socal_state_t *socalp, cqe_t *cqe, caddr_t payload) argument
[all...]
/illumos-gate/usr/src/uts/common/io/fibre-channel/fca/emlxs/
H A Demlxs_sli4.c140 CQE_ASYNC_t *cqe);
142 CQE_ASYNC_t *cqe);
4962 emlxs_sli4_process_async_event(emlxs_hba_t *hba, CQE_ASYNC_t *cqe) argument
4968 if (hba->link_event_tag == cqe->un.link.event_tag) {
4970 } else if (hba->link_event_tag + 1 < cqe->un.link.event_tag) {
4973 hba->link_event_tag = cqe->un.link.event_tag;
4975 switch (cqe->event_code) {
4979 switch (cqe->un.link.link_status) {
4984 cqe->valid, cqe
5188 emlxs_sli4_process_mbox_event(emlxs_hba_t *hba, CQE_MBOX_t *cqe) argument
5407 emlxs_CQE_to_IOCB(emlxs_hba_t *hba, CQE_CmplWQ_t *cqe, emlxs_buf_t *sbp) argument
5519 CQE_CmplWQ_t cqe; local
5602 emlxs_sli4_process_oor_wqe_cmpl(emlxs_hba_t *hba, CQ_DESC_t *cq, CQE_CmplWQ_t *cqe) argument
5626 emlxs_sli4_process_wqe_cmpl(emlxs_hba_t *hba, CQ_DESC_t *cq, CQE_CmplWQ_t *cqe) argument
5753 emlxs_sli4_process_release_wqe(emlxs_hba_t *hba, CQ_DESC_t *cq, CQE_RelWQ_t *cqe) argument
5917 emlxs_sli4_process_unsol_rcv(emlxs_hba_t *hba, CQ_DESC_t *cq, CQE_UnsolRcv_t *cqe) argument
6741 emlxs_sli4_process_xri_aborted(emlxs_hba_t *hba, CQ_DESC_t *cq, CQE_XRI_Abort_t *cqe) argument
6786 CQE_u *cqe; local
9029 emlxs_sli4_handle_fcoe_link_event(emlxs_hba_t *hba, CQE_ASYNC_t *cqe) argument
9080 emlxs_sli4_handle_fc_link_att(emlxs_hba_t *hba, CQE_ASYNC_t *cqe) argument
[all...]

Completed in 643 milliseconds

12