/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License (the "License").
* You may not use this file except in compliance with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright (c) 2008, 2010, Oracle and/or its affiliates. All rights reserved.
* Copyright (c) 2014, Joyent, Inc. All rights reserved.
*/
/*
* - General Introduction:
*
* This file contains the implementation of the MAC client kernel
* API and related code. The MAC client API allows a kernel module
* to gain access to a MAC instance (physical NIC, link aggregation, etc).
* It allows a MAC client to associate itself with a MAC address,
* VLANs, callback functions for data traffic and for promiscuous mode.
* The MAC client API is also used to specify the properties associated
* with a MAC client, such as bandwidth limits, priority, CPUS, etc.
* These properties are further used to determine the hardware resources
* to allocate to the various MAC clients.
*
* - Primary MAC clients:
*
* The MAC client API refers to "primary MAC clients". A primary MAC
* client is a client which "owns" the primary MAC address of
* the underlying MAC instance. The primary MAC address is called out
* since it is associated with specific semantics: the primary MAC
* address is the MAC address which is assigned to the IP interface
* when it is plumbed, and the primary MAC address is assigned
* to VLAN data-links. The primary address of a MAC instance can
* also change dynamically from under the MAC client, for example
* as a result of a change of state of a link aggregation. In that
* case the MAC layer automatically updates all data-structures which
* refer to the current value of the primary MAC address. Typical
* primary MAC clients are dls, aggr, and xnb. A typical non-primary
* MAC client is the vnic driver.
*
* - Virtual Switching:
*
* The MAC layer implements a virtual switch between the MAC clients
* (primary and non-primary) defined on top of the same underlying
* NIC (physical, link aggregation, etc). The virtual switch is
* VLAN-aware, i.e. it allows multiple MAC clients to be member
* of one or more VLANs, and the virtual switch will distribute
* multicast tagged packets only to the member of the corresponding
* VLANs.
*
* - Upper vs Lower MAC:
*
* Creating a VNIC on top of a MAC instance effectively causes
* two MAC instances to be layered on top of each other, one for
* the VNIC(s), one for the underlying MAC instance (physical NIC,
* link aggregation, etc). In the code below we refer to the
* underlying NIC as the "lower MAC", and we refer to VNICs as
* the "upper MAC".
*
* - Pass-through for VNICs:
*
* When VNICs are created on top of an underlying MAC, this causes
* a layering of two MAC instances. Since the lower MAC already
* does the switching and demultiplexing to its MAC clients, the
* upper MAC would simply have to pass packets to the layer below
* or above it, which would introduce overhead. In order to avoid
* this overhead, the MAC layer implements a pass-through mechanism
* for VNICs. When a VNIC opens the lower MAC instance, it saves
* the MAC client handle it optains from the MAC layer. When a MAC
* client opens a VNIC (upper MAC), the MAC layer detects that
* the MAC being opened is a VNIC, and gets the MAC client handle
* that the VNIC driver obtained from the lower MAC. This exchange
* is done through a private capability between the MAC layer
* and the VNIC driver. The upper MAC then returns that handle
* directly to its MAC client. Any operation done by the upper
* MAC client is now done on the lower MAC client handle, which
* allows the VNIC driver to be completely bypassed for the
* performance sensitive data-path.
*
* - Secondary MACs for VNICs:
*
* VNICs support multiple upper mac clients to enable support for
* multiple MAC addresses on the VNIC. When the VNIC is created the
* initial mac client is the primary upper mac. Any additional mac
* clients are secondary macs. These are kept in sync with the primary
* (for things such as the rx function and resource control settings)
* using the same private capability interface between the MAC layer
* and the VNIC layer.
*
*/
#include <sys/types.h>
#include <sys/conf.h>
#include <sys/id_space.h>
#include <sys/esunddi.h>
#include <sys/stat.h>
#include <sys/mkdev.h>
#include <sys/stream.h>
#include <sys/strsun.h>
#include <sys/strsubr.h>
#include <sys/dlpi.h>
#include <sys/modhash.h>
#include <sys/mac_impl.h>
#include <sys/mac_client_impl.h>
#include <sys/mac_soft_ring.h>
#include <sys/mac_stat.h>
#include <sys/dls.h>
#include <sys/dld.h>
#include <sys/modctl.h>
#include <sys/fs/dv_node.h>
#include <sys/thread.h>
#include <sys/proc.h>
#include <sys/callb.h>
#include <sys/cpuvar.h>
#include <sys/atomic.h>
#include <sys/sdt.h>
#include <sys/mac_flow.h>
#include <sys/ddi_intr_impl.h>
#include <sys/disp.h>
#include <sys/sdt.h>
#include <sys/vnic.h>
#include <sys/vnic_impl.h>
#include <sys/vlan.h>
#include <inet/ip.h>
#include <inet/ip6.h>
#include <sys/exacct.h>
#include <sys/exacct_impl.h>
#include <inet/nd.h>
#include <sys/ethernet.h>
kmem_cache_t *mac_client_impl_cache;
kmem_cache_t *mac_promisc_impl_cache;
static boolean_t mac_client_single_rcvr(mac_client_impl_t *);
static flow_entry_t *mac_client_swap_mciflent(mac_client_impl_t *);
static flow_entry_t *mac_client_get_flow(mac_client_impl_t *,
mac_unicast_impl_t *);
static void mac_client_remove_flow_from_list(mac_client_impl_t *,
flow_entry_t *);
static void mac_client_add_to_flow_list(mac_client_impl_t *, flow_entry_t *);
static void mac_rename_flow_names(mac_client_impl_t *, const char *);
static void mac_virtual_link_update(mac_impl_t *);
static int mac_client_datapath_setup(mac_client_impl_t *, uint16_t,
uint8_t *, mac_resource_props_t *, boolean_t, mac_unicast_impl_t *);
static void mac_client_datapath_teardown(mac_client_handle_t,
mac_unicast_impl_t *, flow_entry_t *);
static int mac_resource_ctl_set(mac_client_handle_t, mac_resource_props_t *);
/* ARGSUSED */
static int
i_mac_client_impl_ctor(void *buf, void *arg, int kmflag)
{
int i;
mac_client_impl_t *mcip = buf;
bzero(buf, MAC_CLIENT_IMPL_SIZE);
mutex_init(&mcip->mci_tx_cb_lock, NULL, MUTEX_DRIVER, NULL);
mcip->mci_tx_notify_cb_info.mcbi_lockp = &mcip->mci_tx_cb_lock;
ASSERT(mac_tx_percpu_cnt >= 0);
for (i = 0; i <= mac_tx_percpu_cnt; i++) {
mutex_init(&mcip->mci_tx_pcpu[i].pcpu_tx_lock, NULL,
MUTEX_DRIVER, NULL);
}
cv_init(&mcip->mci_tx_cv, NULL, CV_DRIVER, NULL);
return (0);
}
/* ARGSUSED */
static void
i_mac_client_impl_dtor(void *buf, void *arg)
{
int i;
mac_client_impl_t *mcip = buf;
ASSERT(mcip->mci_promisc_list == NULL);
ASSERT(mcip->mci_unicast_list == NULL);
ASSERT(mcip->mci_state_flags == 0);
ASSERT(mcip->mci_tx_flag == 0);
mutex_destroy(&mcip->mci_tx_cb_lock);
ASSERT(mac_tx_percpu_cnt >= 0);
for (i = 0; i <= mac_tx_percpu_cnt; i++) {
ASSERT(mcip->mci_tx_pcpu[i].pcpu_tx_refcnt == 0);
mutex_destroy(&mcip->mci_tx_pcpu[i].pcpu_tx_lock);
}
cv_destroy(&mcip->mci_tx_cv);
}
/* ARGSUSED */
static int
i_mac_promisc_impl_ctor(void *buf, void *arg, int kmflag)
{
mac_promisc_impl_t *mpip = buf;
bzero(buf, sizeof (mac_promisc_impl_t));
mpip->mpi_mci_link.mcb_objp = buf;
mpip->mpi_mci_link.mcb_objsize = sizeof (mac_promisc_impl_t);
mpip->mpi_mi_link.mcb_objp = buf;
mpip->mpi_mi_link.mcb_objsize = sizeof (mac_promisc_impl_t);
return (0);
}
/* ARGSUSED */
static void
i_mac_promisc_impl_dtor(void *buf, void *arg)
{
mac_promisc_impl_t *mpip = buf;
ASSERT(mpip->mpi_mci_link.mcb_objp != NULL);
ASSERT(mpip->mpi_mci_link.mcb_objsize == sizeof (mac_promisc_impl_t));
ASSERT(mpip->mpi_mi_link.mcb_objp == mpip->mpi_mci_link.mcb_objp);
ASSERT(mpip->mpi_mi_link.mcb_objsize == sizeof (mac_promisc_impl_t));
mpip->mpi_mci_link.mcb_objp = NULL;
mpip->mpi_mci_link.mcb_objsize = 0;
mpip->mpi_mi_link.mcb_objp = NULL;
mpip->mpi_mi_link.mcb_objsize = 0;
ASSERT(mpip->mpi_mci_link.mcb_flags == 0);
mpip->mpi_mci_link.mcb_objsize = 0;
}
void
mac_client_init(void)
{
ASSERT(mac_tx_percpu_cnt >= 0);
mac_client_impl_cache = kmem_cache_create("mac_client_impl_cache",
MAC_CLIENT_IMPL_SIZE, 0, i_mac_client_impl_ctor,
i_mac_client_impl_dtor, NULL, NULL, NULL, 0);
ASSERT(mac_client_impl_cache != NULL);
mac_promisc_impl_cache = kmem_cache_create("mac_promisc_impl_cache",
sizeof (mac_promisc_impl_t), 0, i_mac_promisc_impl_ctor,
i_mac_promisc_impl_dtor, NULL, NULL, NULL, 0);
ASSERT(mac_promisc_impl_cache != NULL);
}
void
mac_client_fini(void)
{
kmem_cache_destroy(mac_client_impl_cache);
kmem_cache_destroy(mac_promisc_impl_cache);
}
/*
* Return the lower MAC client handle from the VNIC driver for the
* specified VNIC MAC instance.
*/
mac_client_impl_t *
mac_vnic_lower(mac_impl_t *mip)
{
mac_capab_vnic_t cap;
mac_client_impl_t *mcip;
VERIFY(i_mac_capab_get((mac_handle_t)mip, MAC_CAPAB_VNIC, &cap));
mcip = cap.mcv_mac_client_handle(cap.mcv_arg);
return (mcip);
}
/*
* Update the secondary macs
*/
void
mac_vnic_secondary_update(mac_impl_t *mip)
{
mac_capab_vnic_t cap;
VERIFY(i_mac_capab_get((mac_handle_t)mip, MAC_CAPAB_VNIC, &cap));
cap.mcv_mac_secondary_update(cap.mcv_arg);
}
/*
* Return the MAC client handle of the primary MAC client for the
* specified MAC instance, or NULL otherwise.
*/
mac_client_impl_t *
mac_primary_client_handle(mac_impl_t *mip)
{
mac_client_impl_t *mcip;
if (mip->mi_state_flags & MIS_IS_VNIC)
return (mac_vnic_lower(mip));
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
for (mcip = mip->mi_clients_list; mcip != NULL;
mcip = mcip->mci_client_next) {
if (MCIP_DATAPATH_SETUP(mcip) && mac_is_primary_client(mcip))
return (mcip);
}
return (NULL);
}
/*
* Open a MAC specified by its MAC name.
*/
int
mac_open(const char *macname, mac_handle_t *mhp)
{
mac_impl_t *mip;
int err;
/*
* Look up its entry in the global hash table.
*/
if ((err = mac_hold(macname, &mip)) != 0)
return (err);
/*
* Hold the dip associated to the MAC to prevent it from being
* detached. For a softmac, its underlying dip is held by the
* mi_open() callback.
*
* This is done to be more tolerant with some defective drivers,
* which incorrectly handle mac_unregister() failure in their
* xxx_detach() routine. For example, some drivers ignore the
* failure of mac_unregister() and free all resources that
* that are needed for data transmition.
*/
e_ddi_hold_devi(mip->mi_dip);
if (!(mip->mi_callbacks->mc_callbacks & MC_OPEN)) {
*mhp = (mac_handle_t)mip;
return (0);
}
/*
* The mac perimeter is used in both mac_open and mac_close by the
* framework to single thread the MC_OPEN/MC_CLOSE of drivers.
*/
i_mac_perim_enter(mip);
mip->mi_oref++;
if (mip->mi_oref != 1 || ((err = mip->mi_open(mip->mi_driver)) == 0)) {
*mhp = (mac_handle_t)mip;
i_mac_perim_exit(mip);
return (0);
}
mip->mi_oref--;
ddi_release_devi(mip->mi_dip);
mac_rele(mip);
i_mac_perim_exit(mip);
return (err);
}
/*
* Open a MAC specified by its linkid.
*/
int
mac_open_by_linkid(datalink_id_t linkid, mac_handle_t *mhp)
{
dls_dl_handle_t dlh;
int err;
if ((err = dls_devnet_hold_tmp(linkid, &dlh)) != 0)
return (err);
dls_devnet_prop_task_wait(dlh);
err = mac_open(dls_devnet_mac(dlh), mhp);
dls_devnet_rele_tmp(dlh);
return (err);
}
/*
* Open a MAC specified by its link name.
*/
int
mac_open_by_linkname(const char *link, mac_handle_t *mhp)
{
datalink_id_t linkid;
int err;
if ((err = dls_mgmt_get_linkid(link, &linkid)) != 0)
return (err);
return (mac_open_by_linkid(linkid, mhp));
}
/*
* Close the specified MAC.
*/
void
mac_close(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
i_mac_perim_enter(mip);
/*
* The mac perimeter is used in both mac_open and mac_close by the
* framework to single thread the MC_OPEN/MC_CLOSE of drivers.
*/
if (mip->mi_callbacks->mc_callbacks & MC_OPEN) {
ASSERT(mip->mi_oref != 0);
if (--mip->mi_oref == 0) {
if ((mip->mi_callbacks->mc_callbacks & MC_CLOSE))
mip->mi_close(mip->mi_driver);
}
}
i_mac_perim_exit(mip);
ddi_release_devi(mip->mi_dip);
mac_rele(mip);
}
/*
* Misc utility functions to retrieve various information about a MAC
* instance or a MAC client.
*/
const mac_info_t *
mac_info(mac_handle_t mh)
{
return (&((mac_impl_t *)mh)->mi_info);
}
dev_info_t *
mac_devinfo_get(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_dip);
}
void *
mac_driver(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_driver);
}
const char *
mac_name(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_name);
}
int
mac_type(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_type->mt_type);
}
int
mac_nativetype(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_type->mt_nativetype);
}
char *
mac_client_name(mac_client_handle_t mch)
{
return (((mac_client_impl_t *)mch)->mci_name);
}
minor_t
mac_minor(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_minor);
}
/*
* Return the VID associated with a MAC client. This function should
* be called for clients which are associated with only one VID.
*/
uint16_t
mac_client_vid(mac_client_handle_t mch)
{
uint16_t vid = VLAN_ID_NONE;
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
flow_desc_t flow_desc;
if (mcip->mci_nflents == 0)
return (vid);
ASSERT(MCIP_DATAPATH_SETUP(mcip) && mac_client_single_rcvr(mcip));
mac_flow_get_desc(mcip->mci_flent, &flow_desc);
if ((flow_desc.fd_mask & FLOW_LINK_VID) != 0)
vid = flow_desc.fd_vid;
return (vid);
}
/*
* Return whether the specified MAC client corresponds to a VLAN VNIC.
*/
boolean_t
mac_client_is_vlan_vnic(mac_client_handle_t mch)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
return (((mcip->mci_state_flags & MCIS_IS_VNIC) != 0) &&
((mcip->mci_flent->fe_type & FLOW_PRIMARY_MAC) != 0));
}
/*
* Return the link speed associated with the specified MAC client.
*
* The link speed of a MAC client is equal to the smallest value of
* 1) the current link speed of the underlying NIC, or
* 2) the bandwidth limit set for the MAC client.
*
* Note that the bandwidth limit can be higher than the speed
* of the underlying NIC. This is allowed to avoid spurious
* administration action failures or artifically lowering the
* bandwidth limit of a link that may have temporarily lowered
* its link speed due to hardware problem or administrator action.
*/
static uint64_t
mac_client_ifspeed(mac_client_impl_t *mcip)
{
mac_impl_t *mip = mcip->mci_mip;
uint64_t nic_speed;
nic_speed = mac_stat_get((mac_handle_t)mip, MAC_STAT_IFSPEED);
if (nic_speed == 0) {
return (0);
} else {
uint64_t policy_limit = (uint64_t)-1;
if (MCIP_RESOURCE_PROPS_MASK(mcip) & MRP_MAXBW)
policy_limit = MCIP_RESOURCE_PROPS_MAXBW(mcip);
return (MIN(policy_limit, nic_speed));
}
}
/*
* Return the link state of the specified client. If here are more
* than one clients of the underying mac_impl_t, the link state
* will always be UP regardless of the link state of the underlying
* mac_impl_t. This is needed to allow the MAC clients to continue
* to communicate with each other even when the physical link of
* their mac_impl_t is down.
*/
static uint64_t
mac_client_link_state(mac_client_impl_t *mcip)
{
mac_impl_t *mip = mcip->mci_mip;
uint16_t vid;
mac_client_impl_t *mci_list;
mac_unicast_impl_t *mui_list, *oth_mui_list;
/*
* Returns LINK_STATE_UP if there are other MAC clients defined on
* mac_impl_t which share same VLAN ID as that of mcip. Note that
* if 'mcip' has more than one VID's then we match ANY one of the
* VID's with other MAC client's VID's and return LINK_STATE_UP.
*/
rw_enter(&mcip->mci_rw_lock, RW_READER);
for (mui_list = mcip->mci_unicast_list; mui_list != NULL;
mui_list = mui_list->mui_next) {
vid = mui_list->mui_vid;
for (mci_list = mip->mi_clients_list; mci_list != NULL;
mci_list = mci_list->mci_client_next) {
if (mci_list == mcip)
continue;
for (oth_mui_list = mci_list->mci_unicast_list;
oth_mui_list != NULL; oth_mui_list = oth_mui_list->
mui_next) {
if (vid == oth_mui_list->mui_vid) {
rw_exit(&mcip->mci_rw_lock);
return (LINK_STATE_UP);
}
}
}
}
rw_exit(&mcip->mci_rw_lock);
return (mac_stat_get((mac_handle_t)mip, MAC_STAT_LINK_STATE));
}
/*
* These statistics are consumed by dladm show-link -s <vnic>,
* dladm show-vnic -s and netstat. With the introduction of dlstat,
* dladm show-link -s and dladm show-vnic -s witll be EOL'ed while
* netstat will consume from kstats introduced for dlstat. This code
* will be removed at that time.
*/
/*
* Return the statistics of a MAC client. These statistics are different
* then the statistics of the underlying MAC which are returned by
* mac_stat_get().
*
* Note that for things based on the tx and rx stats, mac will end up clobbering
* those stats when the underlying set of rings in the srs changes. As such, we
* need to source not only the current set, but also the historical set when
* returning to the client, lest our counters appear to go backwards.
*/
uint64_t
mac_client_stat_get(mac_client_handle_t mch, uint_t stat)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent = mcip->mci_flent;
mac_soft_ring_set_t *mac_srs;
mac_rx_stats_t *mac_rx_stat, *old_rx_stat;
mac_tx_stats_t *mac_tx_stat, *old_tx_stat;
int i;
uint64_t val = 0;
mac_srs = (mac_soft_ring_set_t *)(flent->fe_tx_srs);
mac_tx_stat = &mac_srs->srs_tx.st_stat;
old_rx_stat = &mcip->mci_misc_stat.mms_defunctrxlanestats;
old_tx_stat = &mcip->mci_misc_stat.mms_defuncttxlanestats;
switch (stat) {
case MAC_STAT_LINK_STATE:
val = mac_client_link_state(mcip);
break;
case MAC_STAT_LINK_UP:
val = (mac_client_link_state(mcip) == LINK_STATE_UP);
break;
case MAC_STAT_PROMISC:
val = mac_stat_get((mac_handle_t)mip, MAC_STAT_PROMISC);
break;
case MAC_STAT_LOWLINK_STATE:
val = mac_stat_get((mac_handle_t)mip, MAC_STAT_LOWLINK_STATE);
break;
case MAC_STAT_IFSPEED:
val = mac_client_ifspeed(mcip);
break;
case MAC_STAT_MULTIRCV:
val = mcip->mci_misc_stat.mms_multircv;
break;
case MAC_STAT_BRDCSTRCV:
val = mcip->mci_misc_stat.mms_brdcstrcv;
break;
case MAC_STAT_MULTIXMT:
val = mcip->mci_misc_stat.mms_multixmt;
break;
case MAC_STAT_BRDCSTXMT:
val = mcip->mci_misc_stat.mms_brdcstxmt;
break;
case MAC_STAT_OBYTES:
val = mac_tx_stat->mts_obytes;
val += old_tx_stat->mts_obytes;
break;
case MAC_STAT_OPACKETS:
val = mac_tx_stat->mts_opackets;
val += old_tx_stat->mts_opackets;
break;
case MAC_STAT_OERRORS:
val = mac_tx_stat->mts_oerrors;
val += old_tx_stat->mts_oerrors;
break;
case MAC_STAT_IPACKETS:
for (i = 0; i < flent->fe_rx_srs_cnt; i++) {
mac_srs = (mac_soft_ring_set_t *)flent->fe_rx_srs[i];
mac_rx_stat = &mac_srs->srs_rx.sr_stat;
val += mac_rx_stat->mrs_intrcnt +
mac_rx_stat->mrs_pollcnt + mac_rx_stat->mrs_lclcnt;
}
val += old_rx_stat->mrs_intrcnt + old_rx_stat->mrs_pollcnt +
old_rx_stat->mrs_lclcnt;
break;
case MAC_STAT_RBYTES:
for (i = 0; i < flent->fe_rx_srs_cnt; i++) {
mac_srs = (mac_soft_ring_set_t *)flent->fe_rx_srs[i];
mac_rx_stat = &mac_srs->srs_rx.sr_stat;
val += mac_rx_stat->mrs_intrbytes +
mac_rx_stat->mrs_pollbytes +
mac_rx_stat->mrs_lclbytes;
}
val += old_rx_stat->mrs_intrbytes + old_rx_stat->mrs_pollbytes +
old_rx_stat->mrs_lclbytes;
break;
case MAC_STAT_IERRORS:
for (i = 0; i < flent->fe_rx_srs_cnt; i++) {
mac_srs = (mac_soft_ring_set_t *)flent->fe_rx_srs[i];
mac_rx_stat = &mac_srs->srs_rx.sr_stat;
val += mac_rx_stat->mrs_ierrors;
}
val += old_rx_stat->mrs_ierrors;
break;
default:
val = mac_driver_stat_default(mip, stat);
break;
}
return (val);
}
/*
* Return the statistics of the specified MAC instance.
*/
uint64_t
mac_stat_get(mac_handle_t mh, uint_t stat)
{
mac_impl_t *mip = (mac_impl_t *)mh;
uint64_t val;
int ret;
/*
* The range of stat determines where it is maintained. Stat
* values from 0 up to (but not including) MAC_STAT_MIN are
* mainteined by the mac module itself. Everything else is
* maintained by the driver.
*
* If the mac_impl_t being queried corresponds to a VNIC,
* the stats need to be queried from the lower MAC client
* corresponding to the VNIC. (The mac_link_update()
* invoked by the driver to the lower MAC causes the *lower
* MAC* to update its mi_linkstate, and send a notification
* to its MAC clients. Due to the VNIC passthrough,
* these notifications are sent to the upper MAC clients
* of the VNIC directly, and the upper mac_impl_t of the VNIC
* does not have a valid mi_linkstate.
*/
if (stat < MAC_STAT_MIN && !(mip->mi_state_flags & MIS_IS_VNIC)) {
/* these stats are maintained by the mac module itself */
switch (stat) {
case MAC_STAT_LINK_STATE:
return (mip->mi_linkstate);
case MAC_STAT_LINK_UP:
return (mip->mi_linkstate == LINK_STATE_UP);
case MAC_STAT_PROMISC:
return (mip->mi_devpromisc != 0);
case MAC_STAT_LOWLINK_STATE:
return (mip->mi_lowlinkstate);
default:
ASSERT(B_FALSE);
}
}
/*
* Call the driver to get the given statistic.
*/
ret = mip->mi_getstat(mip->mi_driver, stat, &val);
if (ret != 0) {
/*
* The driver doesn't support this statistic. Get the
* statistic's default value.
*/
val = mac_driver_stat_default(mip, stat);
}
return (val);
}
/*
* Query hardware rx ring corresponding to the pseudo ring.
*/
uint64_t
mac_pseudo_rx_ring_stat_get(mac_ring_handle_t handle, uint_t stat)
{
return (mac_rx_ring_stat_get(handle, stat));
}
/*
* Query hardware tx ring corresponding to the pseudo ring.
*/
uint64_t
mac_pseudo_tx_ring_stat_get(mac_ring_handle_t handle, uint_t stat)
{
return (mac_tx_ring_stat_get(handle, stat));
}
/*
* Utility function which returns the VID associated with a flow entry.
*/
uint16_t
i_mac_flow_vid(flow_entry_t *flent)
{
flow_desc_t flow_desc;
mac_flow_get_desc(flent, &flow_desc);
if ((flow_desc.fd_mask & FLOW_LINK_VID) != 0)
return (flow_desc.fd_vid);
return (VLAN_ID_NONE);
}
/*
* Verify the validity of the specified unicast MAC address. Returns B_TRUE
* if the address is valid, B_FALSE otherwise (multicast address, or incorrect
* length.
*/
boolean_t
mac_unicst_verify(mac_handle_t mh, const uint8_t *addr, uint_t len)
{
mac_impl_t *mip = (mac_impl_t *)mh;
/*
* Verify the address. No lock is needed since mi_type and plugin
* details don't change after mac_register().
*/
if ((len != mip->mi_type->mt_addr_length) ||
(mip->mi_type->mt_ops.mtops_unicst_verify(addr,
mip->mi_pdata)) != 0) {
return (B_FALSE);
} else {
return (B_TRUE);
}
}
void
mac_sdu_get(mac_handle_t mh, uint_t *min_sdu, uint_t *max_sdu)
{
mac_impl_t *mip = (mac_impl_t *)mh;
if (min_sdu != NULL)
*min_sdu = mip->mi_sdu_min;
if (max_sdu != NULL)
*max_sdu = mip->mi_sdu_max;
}
void
mac_sdu_get2(mac_handle_t mh, uint_t *min_sdu, uint_t *max_sdu,
uint_t *multicast_sdu)
{
mac_impl_t *mip = (mac_impl_t *)mh;
if (min_sdu != NULL)
*min_sdu = mip->mi_sdu_min;
if (max_sdu != NULL)
*max_sdu = mip->mi_sdu_max;
if (multicast_sdu != NULL)
*multicast_sdu = mip->mi_sdu_multicast;
}
/*
* Update the MAC unicast address of the specified client's flows. Currently
* only one unicast MAC unicast address is allowed per client.
*/
static void
mac_unicast_update_client_flow(mac_client_impl_t *mcip)
{
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent = mcip->mci_flent;
mac_address_t *map = mcip->mci_unicast;
flow_desc_t flow_desc;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
ASSERT(flent != NULL);
mac_flow_get_desc(flent, &flow_desc);
ASSERT(flow_desc.fd_mask & FLOW_LINK_DST);
bcopy(map->ma_addr, flow_desc.fd_dst_mac, map->ma_len);
mac_flow_set_desc(flent, &flow_desc);
/*
* The v6 local and SLAAC addrs (used by mac protection) need to be
* regenerated because our mac address has changed.
*/
mac_protect_update_mac_token(mcip);
/*
* A MAC client could have one MAC address but multiple
* VLANs. In that case update the flow entries corresponding
* to all VLANs of the MAC client.
*/
for (flent = mcip->mci_flent_list; flent != NULL;
flent = flent->fe_client_next) {
mac_flow_get_desc(flent, &flow_desc);
if (!(flent->fe_type & FLOW_PRIMARY_MAC ||
flent->fe_type & FLOW_VNIC_MAC))
continue;
bcopy(map->ma_addr, flow_desc.fd_dst_mac, map->ma_len);
mac_flow_set_desc(flent, &flow_desc);
}
}
/*
* Update all clients that share the same unicast address.
*/
void
mac_unicast_update_clients(mac_impl_t *mip, mac_address_t *map)
{
mac_client_impl_t *mcip;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
/*
* Find all clients that share the same unicast MAC address and update
* them appropriately.
*/
for (mcip = mip->mi_clients_list; mcip != NULL;
mcip = mcip->mci_client_next) {
/*
* Ignore clients that don't share this MAC address.
*/
if (map != mcip->mci_unicast)
continue;
/*
* Update those clients with same old unicast MAC address.
*/
mac_unicast_update_client_flow(mcip);
}
}
/*
* Update the unicast MAC address of the specified VNIC MAC client.
*
* Check whether the operation is valid. Any of following cases should fail:
*
* 1. It's a VLAN type of VNIC.
* 2. The new value is current "primary" MAC address.
* 3. The current MAC address is shared with other clients.
* 4. The new MAC address has been used. This case will be valid when
* client migration is fully supported.
*/
int
mac_vnic_unicast_set(mac_client_handle_t mch, const uint8_t *addr)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
mac_address_t *map = mcip->mci_unicast;
int err;
ASSERT(!(mip->mi_state_flags & MIS_IS_VNIC));
ASSERT(mcip->mci_state_flags & MCIS_IS_VNIC);
ASSERT(mcip->mci_flags != MAC_CLIENT_FLAGS_PRIMARY);
i_mac_perim_enter(mip);
/*
* If this is a VLAN type of VNIC, it's using "primary" MAC address
* of the underlying interface. Must fail here. Refer to case 1 above.
*/
if (bcmp(map->ma_addr, mip->mi_addr, map->ma_len) == 0) {
i_mac_perim_exit(mip);
return (ENOTSUP);
}
/*
* If the new address is the "primary" one, must fail. Refer to
* case 2 above.
*/
if (bcmp(addr, mip->mi_addr, map->ma_len) == 0) {
i_mac_perim_exit(mip);
return (EACCES);
}
/*
* If the address is shared by multiple clients, must fail. Refer
* to case 3 above.
*/
if (mac_check_macaddr_shared(map)) {
i_mac_perim_exit(mip);
return (EBUSY);
}
/*
* If the new address has been used, must fail for now. Refer to
* case 4 above.
*/
if (mac_find_macaddr(mip, (uint8_t *)addr) != NULL) {
i_mac_perim_exit(mip);
return (ENOTSUP);
}
/*
* Update the MAC address.
*/
err = mac_update_macaddr(map, (uint8_t *)addr);
if (err != 0) {
i_mac_perim_exit(mip);
return (err);
}
/*
* Update all flows of this MAC client.
*/
mac_unicast_update_client_flow(mcip);
i_mac_perim_exit(mip);
return (0);
}
/*
* Program the new primary unicast address of the specified MAC.
*
* Function mac_update_macaddr() takes care different types of underlying
* MAC. If the underlying MAC is VNIC, the VNIC driver must have registerd
* mi_unicst() entry point, that indirectly calls mac_vnic_unicast_set()
* which will take care of updating the MAC address of the corresponding
* MAC client.
*
* This is the only interface that allow the client to update the "primary"
* MAC address of the underlying MAC. The new value must have not been
* used by other clients.
*/
int
mac_unicast_primary_set(mac_handle_t mh, const uint8_t *addr)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_address_t *map;
int err;
/* verify the address validity */
if (!mac_unicst_verify(mh, addr, mip->mi_type->mt_addr_length))
return (EINVAL);
i_mac_perim_enter(mip);
/*
* If the new value is the same as the current primary address value,
* there's nothing to do.
*/
if (bcmp(addr, mip->mi_addr, mip->mi_type->mt_addr_length) == 0) {
i_mac_perim_exit(mip);
return (0);
}
if (mac_find_macaddr(mip, (uint8_t *)addr) != 0) {
i_mac_perim_exit(mip);
return (EBUSY);
}
map = mac_find_macaddr(mip, mip->mi_addr);
ASSERT(map != NULL);
/*
* Update the MAC address.
*/
if (mip->mi_state_flags & MIS_IS_AGGR) {
mac_capab_aggr_t aggr_cap;
/*
* If the mac is an aggregation, other than the unicast
* addresses programming, aggr must be informed about this
* primary unicst address change to change its mac address
* policy to be user-specified.
*/
ASSERT(map->ma_type == MAC_ADDRESS_TYPE_UNICAST_CLASSIFIED);
VERIFY(i_mac_capab_get(mh, MAC_CAPAB_AGGR, &aggr_cap));
err = aggr_cap.mca_unicst(mip->mi_driver, addr);
if (err == 0)
bcopy(addr, map->ma_addr, map->ma_len);
} else {
err = mac_update_macaddr(map, (uint8_t *)addr);
}
if (err != 0) {
i_mac_perim_exit(mip);
return (err);
}
mac_unicast_update_clients(mip, map);
/*
* Save the new primary MAC address in mac_impl_t.
*/
bcopy(addr, mip->mi_addr, mip->mi_type->mt_addr_length);
i_mac_perim_exit(mip);
if (err == 0)
i_mac_notify(mip, MAC_NOTE_UNICST);
return (err);
}
/*
* Return the current primary MAC address of the specified MAC.
*/
void
mac_unicast_primary_get(mac_handle_t mh, uint8_t *addr)
{
mac_impl_t *mip = (mac_impl_t *)mh;
rw_enter(&mip->mi_rw_lock, RW_READER);
bcopy(mip->mi_addr, addr, mip->mi_type->mt_addr_length);
rw_exit(&mip->mi_rw_lock);
}
/*
* Return the secondary MAC address for the specified handle
*/
void
mac_unicast_secondary_get(mac_client_handle_t mh, uint8_t *addr)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mh;
ASSERT(mcip->mci_unicast != NULL);
bcopy(mcip->mci_unicast->ma_addr, addr, mcip->mci_unicast->ma_len);
}
/*
* Return information about the use of the primary MAC address of the
* specified MAC instance:
*
* - if client_name is non-NULL, it must point to a string of at
* least MAXNAMELEN bytes, and will be set to the name of the MAC
* client which uses the primary MAC address.
*
* - if in_use is non-NULL, used to return whether the primary MAC
* address is currently in use.
*/
void
mac_unicast_primary_info(mac_handle_t mh, char *client_name, boolean_t *in_use)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *cur_client;
if (in_use != NULL)
*in_use = B_FALSE;
if (client_name != NULL)
bzero(client_name, MAXNAMELEN);
/*
* The mi_rw_lock is used to protect threads that don't hold the
* mac perimeter to get a consistent view of the mi_clients_list.
* Threads that modify the list must hold both the mac perimeter and
* mi_rw_lock(RW_WRITER)
*/
rw_enter(&mip->mi_rw_lock, RW_READER);
for (cur_client = mip->mi_clients_list; cur_client != NULL;
cur_client = cur_client->mci_client_next) {
if (mac_is_primary_client(cur_client) ||
(mip->mi_state_flags & MIS_IS_VNIC)) {
rw_exit(&mip->mi_rw_lock);
if (in_use != NULL)
*in_use = B_TRUE;
if (client_name != NULL) {
bcopy(cur_client->mci_name, client_name,
MAXNAMELEN);
}
return;
}
}
rw_exit(&mip->mi_rw_lock);
}
/*
* Return the current destination MAC address of the specified MAC.
*/
boolean_t
mac_dst_get(mac_handle_t mh, uint8_t *addr)
{
mac_impl_t *mip = (mac_impl_t *)mh;
rw_enter(&mip->mi_rw_lock, RW_READER);
if (mip->mi_dstaddr_set)
bcopy(mip->mi_dstaddr, addr, mip->mi_type->mt_addr_length);
rw_exit(&mip->mi_rw_lock);
return (mip->mi_dstaddr_set);
}
/*
* Add the specified MAC client to the list of clients which opened
* the specified MAC.
*/
static void
mac_client_add(mac_client_impl_t *mcip)
{
mac_impl_t *mip = mcip->mci_mip;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
/* add VNIC to the front of the list */
rw_enter(&mip->mi_rw_lock, RW_WRITER);
mcip->mci_client_next = mip->mi_clients_list;
mip->mi_clients_list = mcip;
mip->mi_nclients++;
rw_exit(&mip->mi_rw_lock);
}
/*
* Remove the specified MAC client from the list of clients which opened
* the specified MAC.
*/
static void
mac_client_remove(mac_client_impl_t *mcip)
{
mac_impl_t *mip = mcip->mci_mip;
mac_client_impl_t **prev, *cclient;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
rw_enter(&mip->mi_rw_lock, RW_WRITER);
prev = &mip->mi_clients_list;
cclient = *prev;
while (cclient != NULL && cclient != mcip) {
prev = &cclient->mci_client_next;
cclient = *prev;
}
ASSERT(cclient != NULL);
*prev = cclient->mci_client_next;
mip->mi_nclients--;
rw_exit(&mip->mi_rw_lock);
}
static mac_unicast_impl_t *
mac_client_find_vid(mac_client_impl_t *mcip, uint16_t vid)
{
mac_unicast_impl_t *muip = mcip->mci_unicast_list;
while ((muip != NULL) && (muip->mui_vid != vid))
muip = muip->mui_next;
return (muip);
}
/*
* Return whether the specified (MAC address, VID) tuple is already used by
* one of the MAC clients associated with the specified MAC.
*/
static boolean_t
mac_addr_in_use(mac_impl_t *mip, uint8_t *mac_addr, uint16_t vid)
{
mac_client_impl_t *client;
mac_address_t *map;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
for (client = mip->mi_clients_list; client != NULL;
client = client->mci_client_next) {
/*
* Ignore clients that don't have unicast address.
*/
if (client->mci_unicast_list == NULL)
continue;
map = client->mci_unicast;
if ((bcmp(mac_addr, map->ma_addr, map->ma_len) == 0) &&
(mac_client_find_vid(client, vid) != NULL)) {
return (B_TRUE);
}
}
return (B_FALSE);
}
/*
* Generate a random MAC address. The MAC address prefix is
* stored in the array pointed to by mac_addr, and its length, in bytes,
* is specified by prefix_len. The least significant bits
* after prefix_len bytes are generated, and stored after the prefix
* in the mac_addr array.
*/
int
mac_addr_random(mac_client_handle_t mch, uint_t prefix_len,
uint8_t *mac_addr, mac_diag_t *diag)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
size_t addr_len = mip->mi_type->mt_addr_length;
if (prefix_len >= addr_len) {
*diag = MAC_DIAG_MACPREFIXLEN_INVALID;
return (EINVAL);
}
/* check the prefix value */
if (prefix_len > 0) {
bzero(mac_addr + prefix_len, addr_len - prefix_len);
if (!mac_unicst_verify((mac_handle_t)mip, mac_addr,
addr_len)) {
*diag = MAC_DIAG_MACPREFIX_INVALID;
return (EINVAL);
}
}
/* generate the MAC address */
if (prefix_len < addr_len) {
(void) random_get_pseudo_bytes(mac_addr +
prefix_len, addr_len - prefix_len);
}
*diag = 0;
return (0);
}
/*
* Set the priority range for this MAC client. This will be used to
* determine the absolute priority for the threads created for this
* MAC client using the specified "low", "medium" and "high" level.
* This will also be used for any subflows on this MAC client.
*/
#define MAC_CLIENT_SET_PRIORITY_RANGE(mcip, pri) { \
(mcip)->mci_min_pri = FLOW_MIN_PRIORITY(MINCLSYSPRI, \
MAXCLSYSPRI, (pri)); \
(mcip)->mci_max_pri = FLOW_MAX_PRIORITY(MINCLSYSPRI, \
MAXCLSYSPRI, (mcip)->mci_min_pri); \
}
/*
* MAC client open entry point. Return a new MAC client handle. Each
* MAC client is associated with a name, specified through the 'name'
* argument.
*/
int
mac_client_open(mac_handle_t mh, mac_client_handle_t *mchp, char *name,
uint16_t flags)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *mcip;
int err = 0;
boolean_t share_desired;
flow_entry_t *flent = NULL;
share_desired = (flags & MAC_OPEN_FLAGS_SHARES_DESIRED) != 0;
*mchp = NULL;
i_mac_perim_enter(mip);
if (mip->mi_state_flags & MIS_IS_VNIC) {
/*
* The underlying MAC is a VNIC. Return the MAC client
* handle of the lower MAC which was obtained by
* the VNIC driver when it did its mac_client_open().
*/
mcip = mac_vnic_lower(mip);
/*
* Note that multiple mac clients share the same mcip in
* this case.
*/
if (flags & MAC_OPEN_FLAGS_EXCLUSIVE)
mcip->mci_state_flags |= MCIS_EXCLUSIVE;
if (flags & MAC_OPEN_FLAGS_MULTI_PRIMARY)
mcip->mci_flags |= MAC_CLIENT_FLAGS_MULTI_PRIMARY;
mip->mi_clients_list = mcip;
i_mac_perim_exit(mip);
*mchp = (mac_client_handle_t)mcip;
DTRACE_PROBE2(mac__client__open__nonallocated, mac_impl_t *,
mcip->mci_mip, mac_client_impl_t *, mcip);
return (err);
}
mcip = kmem_cache_alloc(mac_client_impl_cache, KM_SLEEP);
mcip->mci_mip = mip;
mcip->mci_upper_mip = NULL;
mcip->mci_rx_fn = mac_pkt_drop;
mcip->mci_rx_arg = NULL;
mcip->mci_rx_p_fn = NULL;
mcip->mci_rx_p_arg = NULL;
mcip->mci_p_unicast_list = NULL;
mcip->mci_direct_rx_fn = NULL;
mcip->mci_direct_rx_arg = NULL;
mcip->mci_vidcache = MCIP_VIDCACHE_INVALID;
mcip->mci_unicast_list = NULL;
if ((flags & MAC_OPEN_FLAGS_IS_VNIC) != 0)
mcip->mci_state_flags |= MCIS_IS_VNIC;
if ((flags & MAC_OPEN_FLAGS_EXCLUSIVE) != 0)
mcip->mci_state_flags |= MCIS_EXCLUSIVE;
if ((flags & MAC_OPEN_FLAGS_IS_AGGR_PORT) != 0)
mcip->mci_state_flags |= MCIS_IS_AGGR_PORT;
if (mip->mi_state_flags & MIS_IS_AGGR)
mcip->mci_state_flags |= MCIS_IS_AGGR;
if ((flags & MAC_OPEN_FLAGS_USE_DATALINK_NAME) != 0) {
datalink_id_t linkid;
ASSERT(name == NULL);
if ((err = dls_devnet_macname2linkid(mip->mi_name,
&linkid)) != 0) {
goto done;
}
if ((err = dls_mgmt_get_linkinfo(linkid, mcip->mci_name, NULL,
NULL, NULL)) != 0) {
/*
* Use mac name if dlmgmtd is not available.
*/
if (err == EBADF) {
(void) strlcpy(mcip->mci_name, mip->mi_name,
sizeof (mcip->mci_name));
err = 0;
} else {
goto done;
}
}
mcip->mci_state_flags |= MCIS_USE_DATALINK_NAME;
} else {
ASSERT(name != NULL);
if (strlen(name) > MAXNAMELEN) {
err = EINVAL;
goto done;
}
(void) strlcpy(mcip->mci_name, name, sizeof (mcip->mci_name));
}
if (flags & MAC_OPEN_FLAGS_MULTI_PRIMARY)
mcip->mci_flags |= MAC_CLIENT_FLAGS_MULTI_PRIMARY;
if (flags & MAC_OPEN_FLAGS_NO_UNICAST_ADDR)
mcip->mci_state_flags |= MCIS_NO_UNICAST_ADDR;
mac_protect_init(mcip);
/* the subflow table will be created dynamically */
mcip->mci_subflow_tab = NULL;
mcip->mci_misc_stat.mms_multircv = 0;
mcip->mci_misc_stat.mms_brdcstrcv = 0;
mcip->mci_misc_stat.mms_multixmt = 0;
mcip->mci_misc_stat.mms_brdcstxmt = 0;
/* Create an initial flow */
err = mac_flow_create(NULL, NULL, mcip->mci_name, NULL,
mcip->mci_state_flags & MCIS_IS_VNIC ? FLOW_VNIC_MAC :
FLOW_PRIMARY_MAC, &flent);
if (err != 0)
goto done;
mcip->mci_flent = flent;
FLOW_MARK(flent, FE_MC_NO_DATAPATH);
flent->fe_mcip = mcip;
/*
* Place initial creation reference on the flow. This reference
* is released in the corresponding delete action viz.
* mac_unicast_remove after waiting for all transient refs to
* to go away. The wait happens in mac_flow_wait.
*/
FLOW_REFHOLD(flent);
/*
* Do this ahead of the mac_bcast_add() below so that the mi_nclients
* will have the right value for mac_rx_srs_setup().
*/
mac_client_add(mcip);
mcip->mci_share = NULL;
if (share_desired)
i_mac_share_alloc(mcip);
/*
* We will do mimimal datapath setup to allow a MAC client to
* transmit or receive non-unicast packets without waiting
* for mac_unicast_add.
*/
if (mcip->mci_state_flags & MCIS_NO_UNICAST_ADDR) {
if ((err = mac_client_datapath_setup(mcip, VLAN_ID_NONE,
NULL, NULL, B_TRUE, NULL)) != 0) {
goto done;
}
}
DTRACE_PROBE2(mac__client__open__allocated, mac_impl_t *,
mcip->mci_mip, mac_client_impl_t *, mcip);
*mchp = (mac_client_handle_t)mcip;
i_mac_perim_exit(mip);
return (0);
done:
i_mac_perim_exit(mip);
mcip->mci_state_flags = 0;
mcip->mci_tx_flag = 0;
kmem_cache_free(mac_client_impl_cache, mcip);
return (err);
}
/*
* Close the specified MAC client handle.
*/
void
mac_client_close(mac_client_handle_t mch, uint16_t flags)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent;
i_mac_perim_enter(mip);
if (flags & MAC_CLOSE_FLAGS_EXCLUSIVE)
mcip->mci_state_flags &= ~MCIS_EXCLUSIVE;
if ((mcip->mci_state_flags & MCIS_IS_VNIC) &&
!(flags & MAC_CLOSE_FLAGS_IS_VNIC)) {
/*
* This is an upper VNIC client initiated operation.
* The lower MAC client will be closed by the VNIC driver
* when the VNIC is deleted.
*/
i_mac_perim_exit(mip);
return;
}
/* If we have only setup up minimal datapth setup, tear it down */
if (mcip->mci_state_flags & MCIS_NO_UNICAST_ADDR) {
mac_client_datapath_teardown((mac_client_handle_t)mcip, NULL,
mcip->mci_flent);
mcip->mci_state_flags &= ~MCIS_NO_UNICAST_ADDR;
}
/*
* Remove the flent associated with the MAC client
*/
flent = mcip->mci_flent;
mcip->mci_flent = NULL;
FLOW_FINAL_REFRELE(flent);
/*
* MAC clients must remove the unicast addresses and promisc callbacks
* they added before issuing a mac_client_close().
*/
ASSERT(mcip->mci_unicast_list == NULL);
ASSERT(mcip->mci_promisc_list == NULL);
ASSERT(mcip->mci_tx_notify_cb_list == NULL);
i_mac_share_free(mcip);
mac_protect_fini(mcip);
mac_client_remove(mcip);
i_mac_perim_exit(mip);
mcip->mci_subflow_tab = NULL;
mcip->mci_state_flags = 0;
mcip->mci_tx_flag = 0;
kmem_cache_free(mac_client_impl_cache, mch);
}
/*
* Set the rx bypass receive callback.
*/
boolean_t
mac_rx_bypass_set(mac_client_handle_t mch, mac_direct_rx_t rx_fn, void *arg1)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
/*
* If the mac_client is a VLAN, we should not do DLS bypass and
* instead let the packets come up via mac_rx_deliver so the vlan
* header can be stripped.
*/
if (mcip->mci_nvids > 0)
return (B_FALSE);
/*
* These are not accessed directly in the data path, and hence
* don't need any protection
*/
mcip->mci_direct_rx_fn = rx_fn;
mcip->mci_direct_rx_arg = arg1;
return (B_TRUE);
}
/*
* Enable/Disable rx bypass. By default, bypass is assumed to be enabled.
*/
void
mac_rx_bypass_enable(mac_client_handle_t mch)
{
((mac_client_impl_t *)mch)->mci_state_flags &= ~MCIS_RX_BYPASS_DISABLE;
}
void
mac_rx_bypass_disable(mac_client_handle_t mch)
{
((mac_client_impl_t *)mch)->mci_state_flags |= MCIS_RX_BYPASS_DISABLE;
}
/*
* Set the receive callback for the specified MAC client. There can be
* at most one such callback per MAC client.
*/
void
mac_rx_set(mac_client_handle_t mch, mac_rx_t rx_fn, void *arg)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
mac_impl_t *umip = mcip->mci_upper_mip;
/*
* Instead of adding an extra set of locks and refcnts in
* the datapath at the mac client boundary, we temporarily quiesce
* the SRS and related entities. We then change the receive function
* without interference from any receive data thread and then reenable
* the data flow subsequently.
*/
i_mac_perim_enter(mip);
mac_rx_client_quiesce(mch);
mcip->mci_rx_fn = rx_fn;
mcip->mci_rx_arg = arg;
mac_rx_client_restart(mch);
i_mac_perim_exit(mip);
/*
* If we're changing the rx function on the primary mac of a vnic,
* make sure any secondary macs on the vnic are updated as well.
*/
if (umip != NULL) {
ASSERT((umip->mi_state_flags & MIS_IS_VNIC) != 0);
mac_vnic_secondary_update(umip);
}
}
/*
* Reset the receive callback for the specified MAC client.
*/
void
mac_rx_clear(mac_client_handle_t mch)
{
mac_rx_set(mch, mac_pkt_drop, NULL);
}
void
mac_secondary_dup(mac_client_handle_t smch, mac_client_handle_t dmch)
{
mac_client_impl_t *smcip = (mac_client_impl_t *)smch;
mac_client_impl_t *dmcip = (mac_client_impl_t *)dmch;
flow_entry_t *flent = dmcip->mci_flent;
/* This should only be called to setup secondary macs */
ASSERT((flent->fe_type & FLOW_PRIMARY_MAC) == 0);
mac_rx_set(dmch, smcip->mci_rx_fn, smcip->mci_rx_arg);
dmcip->mci_promisc_list = smcip->mci_promisc_list;
/*
* Duplicate the primary mac resources to the secondary.
* Since we already validated the resource controls when setting
* them on the primary, we can ignore errors here.
*/
(void) mac_resource_ctl_set(dmch, MCIP_RESOURCE_PROPS(smcip));
}
/*
* Called when removing a secondary MAC. Currently only clears the promisc_list
* since we share the primary mac's promisc_list.
*/
void
mac_secondary_cleanup(mac_client_handle_t mch)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
flow_entry_t *flent = mcip->mci_flent;
/* This should only be called for secondary macs */
ASSERT((flent->fe_type & FLOW_PRIMARY_MAC) == 0);
mcip->mci_promisc_list = NULL;
}
/*
* Walk the MAC client subflow table and updates their priority values.
*/
static int
mac_update_subflow_priority_cb(flow_entry_t *flent, void *arg)
{
mac_flow_update_priority(arg, flent);
return (0);
}
void
mac_update_subflow_priority(mac_client_impl_t *mcip)
{
(void) mac_flow_walk(mcip->mci_subflow_tab,
mac_update_subflow_priority_cb, mcip);
}
/*
* Modify the TX or RX ring properties. We could either just move around
* rings, i.e add/remove rings given to a client. Or this might cause the
* client to move from hardware based to software or the other way around.
* If we want to reset this property, then we clear the mask, additionally
* if the client was given a non-default group we remove all rings except
* for 1 and give it back to the default group.
*/
int
mac_client_set_rings_prop(mac_client_impl_t *mcip, mac_resource_props_t *mrp,
mac_resource_props_t *tmrp)
{
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent = mcip->mci_flent;
uint8_t *mac_addr;
int err = 0;
mac_group_t *defgrp;
mac_group_t *group;
mac_group_t *ngrp;
mac_resource_props_t *cmrp = MCIP_RESOURCE_PROPS(mcip);
uint_t ringcnt;
boolean_t unspec;
if (mcip->mci_share != NULL)
return (EINVAL);
if (mrp->mrp_mask & MRP_RX_RINGS) {
unspec = mrp->mrp_mask & MRP_RXRINGS_UNSPEC;
group = flent->fe_rx_ring_group;
defgrp = MAC_DEFAULT_RX_GROUP(mip);
mac_addr = flent->fe_flow_desc.fd_dst_mac;
/*
* No resulting change. If we are resetting on a client on
* which there was no rx rings property. For dynamic group
* if we are setting the same number of rings already set.
* For static group if we are requesting a group again.
*/
if (mrp->mrp_mask & MRP_RINGS_RESET) {
if (!(tmrp->mrp_mask & MRP_RX_RINGS))
return (0);
} else {
if (unspec) {
if (tmrp->mrp_mask & MRP_RXRINGS_UNSPEC)
return (0);
} else if (mip->mi_rx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
if ((tmrp->mrp_mask & MRP_RX_RINGS) &&
!(tmrp->mrp_mask & MRP_RXRINGS_UNSPEC) &&
mrp->mrp_nrxrings == tmrp->mrp_nrxrings) {
return (0);
}
}
}
/* Resetting the prop */
if (mrp->mrp_mask & MRP_RINGS_RESET) {
/*
* We will just keep one ring and give others back if
* we are not the primary. For the primary we give
* all the rings in the default group except the
* default ring. If it is a static group, then
* we don't do anything, but clear the MRP_RX_RINGS
* flag.
*/
if (group != defgrp) {
if (mip->mi_rx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
/*
* This group has reserved rings
* that need to be released now,
* so does the group.
*/
MAC_RX_RING_RELEASED(mip,
group->mrg_cur_count);
MAC_RX_GRP_RELEASED(mip);
if ((flent->fe_type &
FLOW_PRIMARY_MAC) != 0) {
if (mip->mi_nactiveclients ==
1) {
(void)
mac_rx_switch_group(
mcip, group,
defgrp);
return (0);
} else {
cmrp->mrp_nrxrings =
group->
mrg_cur_count +
defgrp->
mrg_cur_count - 1;
}
} else {
cmrp->mrp_nrxrings = 1;
}
(void) mac_group_ring_modify(mcip,
group, defgrp);
} else {
/*
* If this is a static group, we
* need to release the group. The
* client will remain in the same
* group till some other client
* needs this group.
*/
MAC_RX_GRP_RELEASED(mip);
}
/* Let check if we can give this an excl group */
} else if (group == defgrp) {
ngrp = mac_reserve_rx_group(mcip, mac_addr,
B_TRUE);
/* Couldn't give it a group, that's fine */
if (ngrp == NULL)
return (0);
/* Switch to H/W */
if (mac_rx_switch_group(mcip, defgrp, ngrp) !=
0) {
mac_stop_group(ngrp);
return (0);
}
}
/*
* If the client is in the default group, we will
* just clear the MRP_RX_RINGS and leave it as
* it rather than look for an exclusive group
* for it.
*/
return (0);
}
if (group == defgrp && ((mrp->mrp_nrxrings > 0) || unspec)) {
ngrp = mac_reserve_rx_group(mcip, mac_addr, B_TRUE);
if (ngrp == NULL)
return (ENOSPC);
/* Switch to H/W */
if (mac_rx_switch_group(mcip, defgrp, ngrp) != 0) {
mac_release_rx_group(mcip, ngrp);
return (ENOSPC);
}
MAC_RX_GRP_RESERVED(mip);
if (mip->mi_rx_group_type == MAC_GROUP_TYPE_DYNAMIC)
MAC_RX_RING_RESERVED(mip, ngrp->mrg_cur_count);
} else if (group != defgrp && !unspec &&
mrp->mrp_nrxrings == 0) {
/* Switch to S/W */
ringcnt = group->mrg_cur_count;
if (mac_rx_switch_group(mcip, group, defgrp) != 0)
return (ENOSPC);
if (tmrp->mrp_mask & MRP_RX_RINGS) {
MAC_RX_GRP_RELEASED(mip);
if (mip->mi_rx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
MAC_RX_RING_RELEASED(mip, ringcnt);
}
}
} else if (group != defgrp && mip->mi_rx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
ringcnt = group->mrg_cur_count;
err = mac_group_ring_modify(mcip, group, defgrp);
if (err != 0)
return (err);
/*
* Update the accounting. If this group
* already had explicitly reserved rings,
* we need to update the rings based on
* the new ring count. If this group
* had not explicitly reserved rings,
* then we just reserve the rings asked for
* and reserve the group.
*/
if (tmrp->mrp_mask & MRP_RX_RINGS) {
if (ringcnt > group->mrg_cur_count) {
MAC_RX_RING_RELEASED(mip,
ringcnt - group->mrg_cur_count);
} else {
MAC_RX_RING_RESERVED(mip,
group->mrg_cur_count - ringcnt);
}
} else {
MAC_RX_RING_RESERVED(mip, group->mrg_cur_count);
MAC_RX_GRP_RESERVED(mip);
}
}
}
if (mrp->mrp_mask & MRP_TX_RINGS) {
unspec = mrp->mrp_mask & MRP_TXRINGS_UNSPEC;
group = flent->fe_tx_ring_group;
defgrp = MAC_DEFAULT_TX_GROUP(mip);
/*
* For static groups we only allow rings=0 or resetting the
* rings property.
*/
if (mrp->mrp_ntxrings > 0 &&
mip->mi_tx_group_type != MAC_GROUP_TYPE_DYNAMIC) {
return (ENOTSUP);
}
if (mrp->mrp_mask & MRP_RINGS_RESET) {
if (!(tmrp->mrp_mask & MRP_TX_RINGS))
return (0);
} else {
if (unspec) {
if (tmrp->mrp_mask & MRP_TXRINGS_UNSPEC)
return (0);
} else if (mip->mi_tx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
if ((tmrp->mrp_mask & MRP_TX_RINGS) &&
!(tmrp->mrp_mask & MRP_TXRINGS_UNSPEC) &&
mrp->mrp_ntxrings == tmrp->mrp_ntxrings) {
return (0);
}
}
}
/* Resetting the prop */
if (mrp->mrp_mask & MRP_RINGS_RESET) {
if (group != defgrp) {
if (mip->mi_tx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
ringcnt = group->mrg_cur_count;
if ((flent->fe_type &
FLOW_PRIMARY_MAC) != 0) {
mac_tx_client_quiesce(
(mac_client_handle_t)
mcip);
mac_tx_switch_group(mcip,
group, defgrp);
mac_tx_client_restart(
(mac_client_handle_t)
mcip);
MAC_TX_GRP_RELEASED(mip);
MAC_TX_RING_RELEASED(mip,
ringcnt);
return (0);
}
cmrp->mrp_ntxrings = 1;
(void) mac_group_ring_modify(mcip,
group, defgrp);
/*
* This group has reserved rings
* that need to be released now.
*/
MAC_TX_RING_RELEASED(mip, ringcnt);
}
/*
* If this is a static group, we
* need to release the group. The
* client will remain in the same
* group till some other client
* needs this group.
*/
MAC_TX_GRP_RELEASED(mip);
} else if (group == defgrp &&
(flent->fe_type & FLOW_PRIMARY_MAC) == 0) {
ngrp = mac_reserve_tx_group(mcip, B_TRUE);
if (ngrp == NULL)
return (0);
mac_tx_client_quiesce(
(mac_client_handle_t)mcip);
mac_tx_switch_group(mcip, defgrp, ngrp);
mac_tx_client_restart(
(mac_client_handle_t)mcip);
}
/*
* If the client is in the default group, we will
* just clear the MRP_TX_RINGS and leave it as
* it rather than look for an exclusive group
* for it.
*/
return (0);
}
/* Switch to H/W */
if (group == defgrp && ((mrp->mrp_ntxrings > 0) || unspec)) {
ngrp = mac_reserve_tx_group(mcip, B_TRUE);
if (ngrp == NULL)
return (ENOSPC);
mac_tx_client_quiesce((mac_client_handle_t)mcip);
mac_tx_switch_group(mcip, defgrp, ngrp);
mac_tx_client_restart((mac_client_handle_t)mcip);
MAC_TX_GRP_RESERVED(mip);
if (mip->mi_tx_group_type == MAC_GROUP_TYPE_DYNAMIC)
MAC_TX_RING_RESERVED(mip, ngrp->mrg_cur_count);
/* Switch to S/W */
} else if (group != defgrp && !unspec &&
mrp->mrp_ntxrings == 0) {
/* Switch to S/W */
ringcnt = group->mrg_cur_count;
mac_tx_client_quiesce((mac_client_handle_t)mcip);
mac_tx_switch_group(mcip, group, defgrp);
mac_tx_client_restart((mac_client_handle_t)mcip);
if (tmrp->mrp_mask & MRP_TX_RINGS) {
MAC_TX_GRP_RELEASED(mip);
if (mip->mi_tx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
MAC_TX_RING_RELEASED(mip, ringcnt);
}
}
} else if (group != defgrp && mip->mi_tx_group_type ==
MAC_GROUP_TYPE_DYNAMIC) {
ringcnt = group->mrg_cur_count;
err = mac_group_ring_modify(mcip, group, defgrp);
if (err != 0)
return (err);
/*
* Update the accounting. If this group
* already had explicitly reserved rings,
* we need to update the rings based on
* the new ring count. If this group
* had not explicitly reserved rings,
* then we just reserve the rings asked for
* and reserve the group.
*/
if (tmrp->mrp_mask & MRP_TX_RINGS) {
if (ringcnt > group->mrg_cur_count) {
MAC_TX_RING_RELEASED(mip,
ringcnt - group->mrg_cur_count);
} else {
MAC_TX_RING_RESERVED(mip,
group->mrg_cur_count - ringcnt);
}
} else {
MAC_TX_RING_RESERVED(mip, group->mrg_cur_count);
MAC_TX_GRP_RESERVED(mip);
}
}
}
return (0);
}
/*
* When the MAC client is being brought up (i.e. we do a unicast_add) we need
* to initialize the cpu and resource control structure in the
* mac_client_impl_t from the mac_impl_t (i.e if there are any cached
* properties before the flow entry for the unicast address was created).
*/
static int
mac_resource_ctl_set(mac_client_handle_t mch, mac_resource_props_t *mrp)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = (mac_impl_t *)mcip->mci_mip;
mac_impl_t *umip = mcip->mci_upper_mip;
int err = 0;
flow_entry_t *flent = mcip->mci_flent;
mac_resource_props_t *omrp, *nmrp = MCIP_RESOURCE_PROPS(mcip);
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
err = mac_validate_props(mcip->mci_state_flags & MCIS_IS_VNIC ?
mcip->mci_upper_mip : mip, mrp);
if (err != 0)
return (err);
/*
* Copy over the existing properties since mac_update_resources
* will modify the client's mrp. Currently, the saved property
* is used to determine the difference between existing and
* modified rings property.
*/
omrp = kmem_zalloc(sizeof (*omrp), KM_SLEEP);
bcopy(nmrp, omrp, sizeof (*omrp));
mac_update_resources(mrp, MCIP_RESOURCE_PROPS(mcip), B_FALSE);
if (MCIP_DATAPATH_SETUP(mcip)) {
/*
* We support rings only for primary client when there are
* multiple clients sharing the same MAC address (e.g. VLAN).
*/
if (mrp->mrp_mask & MRP_RX_RINGS ||
mrp->mrp_mask & MRP_TX_RINGS) {
if ((err = mac_client_set_rings_prop(mcip, mrp,
omrp)) != 0) {
if (omrp->mrp_mask & MRP_RX_RINGS) {
nmrp->mrp_mask |= MRP_RX_RINGS;
nmrp->mrp_nrxrings = omrp->mrp_nrxrings;
} else {
nmrp->mrp_mask &= ~MRP_RX_RINGS;
nmrp->mrp_nrxrings = 0;
}
if (omrp->mrp_mask & MRP_TX_RINGS) {
nmrp->mrp_mask |= MRP_TX_RINGS;
nmrp->mrp_ntxrings = omrp->mrp_ntxrings;
} else {
nmrp->mrp_mask &= ~MRP_TX_RINGS;
nmrp->mrp_ntxrings = 0;
}
if (omrp->mrp_mask & MRP_RXRINGS_UNSPEC)
omrp->mrp_mask |= MRP_RXRINGS_UNSPEC;
else
omrp->mrp_mask &= ~MRP_RXRINGS_UNSPEC;
if (omrp->mrp_mask & MRP_TXRINGS_UNSPEC)
omrp->mrp_mask |= MRP_TXRINGS_UNSPEC;
else
omrp->mrp_mask &= ~MRP_TXRINGS_UNSPEC;
kmem_free(omrp, sizeof (*omrp));
return (err);
}
/*
* If we modified the rings property of the primary
* we need to update the property fields of its
* VLANs as they inherit the primary's properites.
*/
if (mac_is_primary_client(mcip)) {
mac_set_prim_vlan_rings(mip,
MCIP_RESOURCE_PROPS(mcip));
}
}
/*
* We have to set this prior to calling mac_flow_modify.
*/
if (mrp->mrp_mask & MRP_PRIORITY) {
if (mrp->mrp_priority == MPL_RESET) {
MAC_CLIENT_SET_PRIORITY_RANGE(mcip,
MPL_LINK_DEFAULT);
} else {
MAC_CLIENT_SET_PRIORITY_RANGE(mcip,
mrp->mrp_priority);
}
}
mac_flow_modify(mip->mi_flow_tab, flent, mrp);
if (mrp->mrp_mask & MRP_PRIORITY)
mac_update_subflow_priority(mcip);
/* Apply these resource settings to any secondary macs */
if (umip != NULL) {
ASSERT((umip->mi_state_flags & MIS_IS_VNIC) != 0);
mac_vnic_secondary_update(umip);
}
}
kmem_free(omrp, sizeof (*omrp));
return (0);
}
static int
mac_unicast_flow_create(mac_client_impl_t *mcip, uint8_t *mac_addr,
uint16_t vid, boolean_t is_primary, boolean_t first_flow,
flow_entry_t **flent, mac_resource_props_t *mrp)
{
mac_impl_t *mip = (mac_impl_t *)mcip->mci_mip;
flow_desc_t flow_desc;
char flowname[MAXFLOWNAMELEN];
int err;
uint_t flent_flags;
/*
* First unicast address being added, create a new flow
* for that MAC client.
*/
bzero(&flow_desc, sizeof (flow_desc));
ASSERT(mac_addr != NULL ||
(mcip->mci_state_flags & MCIS_NO_UNICAST_ADDR));
if (mac_addr != NULL) {
flow_desc.fd_mac_len = mip->mi_type->mt_addr_length;
bcopy(mac_addr, flow_desc.fd_dst_mac, flow_desc.fd_mac_len);
}
flow_desc.fd_mask = FLOW_LINK_DST;
if (vid != 0) {
flow_desc.fd_vid = vid;
flow_desc.fd_mask |= FLOW_LINK_VID;
}
/*
* XXX-nicolas. For now I'm keeping the FLOW_PRIMARY_MAC
* and FLOW_VNIC. Even though they're a hack inherited
* from the SRS code, we'll keep them for now. They're currently
* consumed by mac_datapath_setup() to create the SRS.
* That code should be eventually moved out of
* mac_datapath_setup() and moved to a mac_srs_create()
* function of some sort to keep things clean.
*
* Also, there's no reason why the SRS for the primary MAC
* client should be different than any other MAC client. Until
* this is cleaned-up, we support only one MAC unicast address
* per client.
*
* We set FLOW_PRIMARY_MAC for the primary MAC address,
* FLOW_VNIC for everything else.
*/
if (is_primary)
flent_flags = FLOW_PRIMARY_MAC;
else
flent_flags = FLOW_VNIC_MAC;
/*
* For the first flow we use the mac client's name - mci_name, for
* subsequent ones we just create a name with the vid. This is
* so that we can add these flows to the same flow table. This is
* fine as the flow name (except for the one with the mac client's
* name) is not visible. When the first flow is removed, we just replace
* its fdesc with another from the list, so we will still retain the
* flent with the MAC client's flow name.
*/
if (first_flow) {
bcopy(mcip->mci_name, flowname, MAXFLOWNAMELEN);
} else {
(void) sprintf(flowname, "%s%u", mcip->mci_name, vid);
flent_flags = FLOW_NO_STATS;
}
if ((err = mac_flow_create(&flow_desc, mrp, flowname, NULL,
flent_flags, flent)) != 0)
return (err);
mac_misc_stat_create(*flent);
FLOW_MARK(*flent, FE_INCIPIENT);
(*flent)->fe_mcip = mcip;
/*
* Place initial creation reference on the flow. This reference
* is released in the corresponding delete action viz.
* mac_unicast_remove after waiting for all transient refs to
* to go away. The wait happens in mac_flow_wait.
* We have already held the reference in mac_client_open().
*/
if (!first_flow)
FLOW_REFHOLD(*flent);
return (0);
}
/* Refresh the multicast grouping for this VID. */
int
mac_client_update_mcast(void *arg, boolean_t add, const uint8_t *addrp)
{
flow_entry_t *flent = arg;
mac_client_impl_t *mcip = flent->fe_mcip;
uint16_t vid;
flow_desc_t flow_desc;
mac_flow_get_desc(flent, &flow_desc);
vid = (flow_desc.fd_mask & FLOW_LINK_VID) != 0 ?
flow_desc.fd_vid : VLAN_ID_NONE;
/*
* We don't call mac_multicast_add()/mac_multicast_remove() as
* we want to add/remove for this specific vid.
*/
if (add) {
return (mac_bcast_add(mcip, addrp, vid,
MAC_ADDRTYPE_MULTICAST));
} else {
mac_bcast_delete(mcip, addrp, vid);
return (0);
}
}
static void
mac_update_single_active_client(mac_impl_t *mip)
{
mac_client_impl_t *client = NULL;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
rw_enter(&mip->mi_rw_lock, RW_WRITER);
if (mip->mi_nactiveclients == 1) {
/*
* Find the one active MAC client from the list of MAC
* clients. The active MAC client has at least one
* unicast address.
*/
for (client = mip->mi_clients_list; client != NULL;
client = client->mci_client_next) {
if (client->mci_unicast_list != NULL)
break;
}
ASSERT(client != NULL);
}
/*
* mi_single_active_client is protected by the MAC impl's read/writer
* lock, which allows mac_rx() to check the value of that pointer
* as a reader.
*/
mip->mi_single_active_client = client;
rw_exit(&mip->mi_rw_lock);
}
/*
* Set up the data path. Called from i_mac_unicast_add after having
* done all the validations including making sure this is an active
* client (i.e that is ready to process packets.)
*/
static int
mac_client_datapath_setup(mac_client_impl_t *mcip, uint16_t vid,
uint8_t *mac_addr, mac_resource_props_t *mrp, boolean_t isprimary,
mac_unicast_impl_t *muip)
{
mac_impl_t *mip = mcip->mci_mip;
boolean_t mac_started = B_FALSE;
boolean_t bcast_added = B_FALSE;
boolean_t nactiveclients_added = B_FALSE;
flow_entry_t *flent;
int err = 0;
boolean_t no_unicast;
no_unicast = mcip->mci_state_flags & MCIS_NO_UNICAST_ADDR;
if ((err = mac_start((mac_handle_t)mip)) != 0)
goto bail;
mac_started = B_TRUE;
/* add the MAC client to the broadcast address group by default */
if (mip->mi_type->mt_brdcst_addr != NULL) {
err = mac_bcast_add(mcip, mip->mi_type->mt_brdcst_addr, vid,
MAC_ADDRTYPE_BROADCAST);
if (err != 0)
goto bail;
bcast_added = B_TRUE;
}
/*
* If this is the first unicast address addition for this
* client, reuse the pre-allocated larval flow entry associated with
* the MAC client.
*/
flent = (mcip->mci_nflents == 0) ? mcip->mci_flent : NULL;
/* We are configuring the unicast flow now */
if (!MCIP_DATAPATH_SETUP(mcip)) {
if (mrp != NULL) {
MAC_CLIENT_SET_PRIORITY_RANGE(mcip,
(mrp->mrp_mask & MRP_PRIORITY) ? mrp->mrp_priority :
MPL_LINK_DEFAULT);
}
if ((err = mac_unicast_flow_create(mcip, mac_addr, vid,
isprimary, B_TRUE, &flent, mrp)) != 0)
goto bail;
mip->mi_nactiveclients++;
nactiveclients_added = B_TRUE;
/*
* This will allocate the RX ring group if possible for the
* flow and program the software classifier as needed.
*/
if ((err = mac_datapath_setup(mcip, flent, SRST_LINK)) != 0)
goto bail;
if (no_unicast)
goto done_setup;
/*
* The unicast MAC address must have been added successfully.
*/
ASSERT(mcip->mci_unicast != NULL);
/*
* Push down the sub-flows that were defined on this link
* hitherto. The flows are added to the active flow table
* and SRS, softrings etc. are created as needed.
*/
mac_link_init_flows((mac_client_handle_t)mcip);
} else {
mac_address_t *map = mcip->mci_unicast;
ASSERT(!no_unicast);
/*
* A unicast flow already exists for that MAC client,
* this flow must be the same mac address but with
* different VID. It has been checked by mac_addr_in_use().
*
* We will use the SRS etc. from the mci_flent. Note that
* We don't need to create kstat for this as except for
* the fdesc, everything will be used from in the 1st flent.
*/
if (bcmp(mac_addr, map->ma_addr, map->ma_len) != 0) {
err = EINVAL;
goto bail;
}
if ((err = mac_unicast_flow_create(mcip, mac_addr, vid,
isprimary, B_FALSE, &flent, NULL)) != 0) {
goto bail;
}
if ((err = mac_flow_add(mip->mi_flow_tab, flent)) != 0) {
FLOW_FINAL_REFRELE(flent);
goto bail;
}
/* update the multicast group for this vid */
mac_client_bcast_refresh(mcip, mac_client_update_mcast,
(void *)flent, B_TRUE);
}
/* populate the shared MAC address */
muip->mui_map = mcip->mci_unicast;
rw_enter(&mcip->mci_rw_lock, RW_WRITER);
muip->mui_next = mcip->mci_unicast_list;
mcip->mci_unicast_list = muip;
rw_exit(&mcip->mci_rw_lock);
done_setup:
/*
* First add the flent to the flow list of this mcip. Then set
* the mip's mi_single_active_client if needed. The Rx path assumes
* that mip->mi_single_active_client will always have an associated
* flent.
*/
mac_client_add_to_flow_list(mcip, flent);
if (nactiveclients_added)
mac_update_single_active_client(mip);
/*
* Trigger a renegotiation of the capabilities when the number of
* active clients changes from 1 to 2, since some of the capabilities
* might have to be disabled. Also send a MAC_NOTE_LINK notification
* to all the MAC clients whenever physical link is DOWN.
*/
if (mip->mi_nactiveclients == 2) {
mac_capab_update((mac_handle_t)mip);
mac_virtual_link_update(mip);
}
/*
* Now that the setup is complete, clear the INCIPIENT flag.
* The flag was set to avoid incoming packets seeing inconsistent
* structures while the setup was in progress. Clear the mci_tx_flag
* by calling mac_tx_client_block. It is possible that
* mac_unicast_remove was called prior to this mac_unicast_add which
* could have set the MCI_TX_QUIESCE flag.
*/
if (flent->fe_rx_ring_group != NULL)
mac_rx_group_unmark(flent->fe_rx_ring_group, MR_INCIPIENT);
FLOW_UNMARK(flent, FE_INCIPIENT);
FLOW_UNMARK(flent, FE_MC_NO_DATAPATH);
mac_tx_client_unblock(mcip);
return (0);
bail:
if (bcast_added)
mac_bcast_delete(mcip, mip->mi_type->mt_brdcst_addr, vid);
if (nactiveclients_added)
mip->mi_nactiveclients--;
if (mac_started)
mac_stop((mac_handle_t)mip);
return (err);
}
/*
* Return the passive primary MAC client, if present. The passive client is
* a stand-by client that has the same unicast address as another that is
* currenly active. Once the active client goes away, the passive client
* becomes active.
*/
static mac_client_impl_t *
mac_get_passive_primary_client(mac_impl_t *mip)
{
mac_client_impl_t *mcip;
for (mcip = mip->mi_clients_list; mcip != NULL;
mcip = mcip->mci_client_next) {
if (mac_is_primary_client(mcip) &&
(mcip->mci_flags & MAC_CLIENT_FLAGS_PASSIVE_PRIMARY) != 0) {
return (mcip);
}
}
return (NULL);
}
/*
* Add a new unicast address to the MAC client.
*
* The MAC address can be specified either by value, or the MAC client
* can specify that it wants to use the primary MAC address of the
* underlying MAC. See the introductory comments at the beginning
* of this file for more more information on primary MAC addresses.
*
* Note also the tuple (MAC address, VID) must be unique
* for the MAC clients defined on top of the same underlying MAC
* instance, unless the MAC_UNICAST_NODUPCHECK is specified.
*
* In no case can a client use the PVID for the MAC, if the MAC has one set.
*/
int
i_mac_unicast_add(mac_client_handle_t mch, uint8_t *mac_addr, uint16_t flags,
mac_unicast_handle_t *mah, uint16_t vid, mac_diag_t *diag)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
int err;
uint_t mac_len = mip->mi_type->mt_addr_length;
boolean_t check_dups = !(flags & MAC_UNICAST_NODUPCHECK);
boolean_t fastpath_disabled = B_FALSE;
boolean_t is_primary = (flags & MAC_UNICAST_PRIMARY);
boolean_t is_unicast_hw = (flags & MAC_UNICAST_HW);
mac_resource_props_t *mrp;
boolean_t passive_client = B_FALSE;
mac_unicast_impl_t *muip;
boolean_t is_vnic_primary =
(flags & MAC_UNICAST_VNIC_PRIMARY);
/* when VID is non-zero, the underlying MAC can not be VNIC */
ASSERT(!((mip->mi_state_flags & MIS_IS_VNIC) && (vid != 0)));
/*
* Can't unicast add if the client asked only for minimal datapath
* setup.
*/
if (mcip->mci_state_flags & MCIS_NO_UNICAST_ADDR)
return (ENOTSUP);
/*
* Check for an attempted use of the current Port VLAN ID, if enabled.
* No client may use it.
*/
if (mip->mi_pvid != 0 && vid == mip->mi_pvid)
return (EBUSY);
/*
* Check whether it's the primary client and flag it.
*/
if (!(mcip->mci_state_flags & MCIS_IS_VNIC) && is_primary && vid == 0)
mcip->mci_flags |= MAC_CLIENT_FLAGS_PRIMARY;
/*
* is_vnic_primary is true when we come here as a VLAN VNIC
* which uses the primary mac client's address but with a non-zero
* VID. In this case the MAC address is not specified by an upper
* MAC client.
*/
if ((mcip->mci_state_flags & MCIS_IS_VNIC) && is_primary &&
!is_vnic_primary) {
/*
* The address is being set by the upper MAC client
* of a VNIC. The MAC address was already set by the
* VNIC driver during VNIC creation.
*
* Note: a VNIC has only one MAC address. We return
* the MAC unicast address handle of the lower MAC client
* corresponding to the VNIC. We allocate a new entry
* which is flagged appropriately, so that mac_unicast_remove()
* doesn't attempt to free the original entry that
* was allocated by the VNIC driver.
*/
ASSERT(mcip->mci_unicast != NULL);
/* Check for VLAN flags, if present */
if ((flags & MAC_UNICAST_TAG_DISABLE) != 0)
mcip->mci_state_flags |= MCIS_TAG_DISABLE;
if ((flags & MAC_UNICAST_STRIP_DISABLE) != 0)
mcip->mci_state_flags |= MCIS_STRIP_DISABLE;
if ((flags & MAC_UNICAST_DISABLE_TX_VID_CHECK) != 0)
mcip->mci_state_flags |= MCIS_DISABLE_TX_VID_CHECK;
/*
* Ensure that the primary unicast address of the VNIC
* is added only once unless we have the
* MAC_CLIENT_FLAGS_MULTI_PRIMARY set (and this is not
* a passive MAC client).
*/
if ((mcip->mci_flags & MAC_CLIENT_FLAGS_VNIC_PRIMARY) != 0) {
if ((mcip->mci_flags &
MAC_CLIENT_FLAGS_MULTI_PRIMARY) == 0 ||
(mcip->mci_flags &
MAC_CLIENT_FLAGS_PASSIVE_PRIMARY) != 0) {
return (EBUSY);
}
mcip->mci_flags |= MAC_CLIENT_FLAGS_PASSIVE_PRIMARY;
passive_client = B_TRUE;
}
mcip->mci_flags |= MAC_CLIENT_FLAGS_VNIC_PRIMARY;
/*
* Create a handle for vid 0.
*/
ASSERT(vid == 0);
muip = kmem_zalloc(sizeof (mac_unicast_impl_t), KM_SLEEP);
muip->mui_vid = vid;
*mah = (mac_unicast_handle_t)muip;
/*
* This will be used by the caller to defer setting the
* rx functions.
*/
if (passive_client)
return (EAGAIN);
return (0);
}
/* primary MAC clients cannot be opened on top of anchor VNICs */
if ((is_vnic_primary || is_primary) &&
i_mac_capab_get((mac_handle_t)mip, MAC_CAPAB_ANCHOR_VNIC, NULL)) {
return (ENXIO);
}
/*
* If this is a VNIC/VLAN, disable softmac fast-path.
*/
if (mcip->mci_state_flags & MCIS_IS_VNIC) {
err = mac_fastpath_disable((mac_handle_t)mip);
if (err != 0)
return (err);
fastpath_disabled = B_TRUE;
}
/*
* Return EBUSY if:
* - there is an exclusively active mac client exists.
* - this is an exclusive active mac client but
* a. there is already active mac clients exist, or
* b. fastpath streams are already plumbed on this legacy device
* - the mac creator has disallowed active mac clients.
*/
if (mip->mi_state_flags & (MIS_EXCLUSIVE|MIS_NO_ACTIVE)) {
if (fastpath_disabled)
mac_fastpath_enable((mac_handle_t)mip);
return (EBUSY);
}
if (mcip->mci_state_flags & MCIS_EXCLUSIVE) {
ASSERT(!fastpath_disabled);
if (mip->mi_nactiveclients != 0)
return (EBUSY);
if ((mip->mi_state_flags & MIS_LEGACY) &&
!(mip->mi_capab_legacy.ml_active_set(mip->mi_driver))) {
return (EBUSY);
}
mip->mi_state_flags |= MIS_EXCLUSIVE;
}
mrp = kmem_zalloc(sizeof (*mrp), KM_SLEEP);
if (is_primary && !(mcip->mci_state_flags & (MCIS_IS_VNIC |
MCIS_IS_AGGR_PORT))) {
/*
* Apply the property cached in the mac_impl_t to the primary
* mac client. If the mac client is a VNIC or an aggregation
* port, its property should be set in the mcip when the
* VNIC/aggr was created.
*/
mac_get_resources((mac_handle_t)mip, mrp);
(void) mac_client_set_resources(mch, mrp);
} else if (mcip->mci_state_flags & MCIS_IS_VNIC) {
/*
* This is a primary VLAN client, we don't support
* specifying rings property for this as it inherits the
* rings property from its MAC.
*/
if (is_vnic_primary) {
mac_resource_props_t *vmrp;
vmrp = MCIP_RESOURCE_PROPS(mcip);
if (vmrp->mrp_mask & MRP_RX_RINGS ||
vmrp->mrp_mask & MRP_TX_RINGS) {
if (fastpath_disabled)
mac_fastpath_enable((mac_handle_t)mip);
kmem_free(mrp, sizeof (*mrp));
return (ENOTSUP);
}
/*
* Additionally we also need to inherit any
* rings property from the MAC.
*/
mac_get_resources((mac_handle_t)mip, mrp);
if (mrp->mrp_mask & MRP_RX_RINGS) {
vmrp->mrp_mask |= MRP_RX_RINGS;
vmrp->mrp_nrxrings = mrp->mrp_nrxrings;
}
if (mrp->mrp_mask & MRP_TX_RINGS) {
vmrp->mrp_mask |= MRP_TX_RINGS;
vmrp->mrp_ntxrings = mrp->mrp_ntxrings;
}
}
bcopy(MCIP_RESOURCE_PROPS(mcip), mrp, sizeof (*mrp));
}
muip = kmem_zalloc(sizeof (mac_unicast_impl_t), KM_SLEEP);
muip->mui_vid = vid;
if (is_primary || is_vnic_primary) {
mac_addr = mip->mi_addr;
} else {
/*
* Verify the validity of the specified MAC addresses value.
*/
if (!mac_unicst_verify((mac_handle_t)mip, mac_addr, mac_len)) {
*diag = MAC_DIAG_MACADDR_INVALID;
err = EINVAL;
goto bail_out;
}
/*
* Make sure that the specified MAC address is different
* than the unicast MAC address of the underlying NIC.
*/
if (check_dups && bcmp(mip->mi_addr, mac_addr, mac_len) == 0) {
*diag = MAC_DIAG_MACADDR_NIC;
err = EINVAL;
goto bail_out;
}
}
/*
* Set the flags here so that if this is a passive client, we
* can return and set it when we call mac_client_datapath_setup
* when this becomes the active client. If we defer to using these
* flags to mac_client_datapath_setup, then for a passive client,
* we'd have to store the flags somewhere (probably fe_flags)
* and then use it.
*/
if (!MCIP_DATAPATH_SETUP(mcip)) {
if (is_unicast_hw) {
/*
* The client requires a hardware MAC address slot
* for that unicast address. Since we support only
* one unicast MAC address per client, flag the
* MAC client itself.
*/
mcip->mci_state_flags |= MCIS_UNICAST_HW;
}
/* Check for VLAN flags, if present */
if ((flags & MAC_UNICAST_TAG_DISABLE) != 0)
mcip->mci_state_flags |= MCIS_TAG_DISABLE;
if ((flags & MAC_UNICAST_STRIP_DISABLE) != 0)
mcip->mci_state_flags |= MCIS_STRIP_DISABLE;
if ((flags & MAC_UNICAST_DISABLE_TX_VID_CHECK) != 0)
mcip->mci_state_flags |= MCIS_DISABLE_TX_VID_CHECK;
} else {
/*
* Assert that the specified flags are consistent with the
* flags specified by previous calls to mac_unicast_add().
*/
ASSERT(((flags & MAC_UNICAST_TAG_DISABLE) != 0 &&
(mcip->mci_state_flags & MCIS_TAG_DISABLE) != 0) ||
((flags & MAC_UNICAST_TAG_DISABLE) == 0 &&
(mcip->mci_state_flags & MCIS_TAG_DISABLE) == 0));
ASSERT(((flags & MAC_UNICAST_STRIP_DISABLE) != 0 &&
(mcip->mci_state_flags & MCIS_STRIP_DISABLE) != 0) ||
((flags & MAC_UNICAST_STRIP_DISABLE) == 0 &&
(mcip->mci_state_flags & MCIS_STRIP_DISABLE) == 0));
ASSERT(((flags & MAC_UNICAST_DISABLE_TX_VID_CHECK) != 0 &&
(mcip->mci_state_flags & MCIS_DISABLE_TX_VID_CHECK) != 0) ||
((flags & MAC_UNICAST_DISABLE_TX_VID_CHECK) == 0 &&
(mcip->mci_state_flags & MCIS_DISABLE_TX_VID_CHECK) == 0));
/*
* Make sure the client is consistent about its requests
* for MAC addresses. I.e. all requests from the clients
* must have the MAC_UNICAST_HW flag set or clear.
*/
if ((mcip->mci_state_flags & MCIS_UNICAST_HW) != 0 &&
!is_unicast_hw ||
(mcip->mci_state_flags & MCIS_UNICAST_HW) == 0 &&
is_unicast_hw) {
err = EINVAL;
goto bail_out;
}
}
/*
* Make sure the MAC address is not already used by
* another MAC client defined on top of the same
* underlying NIC. Unless we have MAC_CLIENT_FLAGS_MULTI_PRIMARY
* set when we allow a passive client to be present which will
* be activated when the currently active client goes away - this
* works only with primary addresses.
*/
if ((check_dups || is_primary || is_vnic_primary) &&
mac_addr_in_use(mip, mac_addr, vid)) {
/*
* Must have set the multiple primary address flag when
* we did a mac_client_open AND this should be a primary
* MAC client AND there should not already be a passive
* primary. If all is true then we let this succeed
* even if the address is a dup.
*/
if ((mcip->mci_flags & MAC_CLIENT_FLAGS_MULTI_PRIMARY) == 0 ||
(mcip->mci_flags & MAC_CLIENT_FLAGS_PRIMARY) == 0 ||
mac_get_passive_primary_client(mip) != NULL) {
*diag = MAC_DIAG_MACADDR_INUSE;
err = EEXIST;
goto bail_out;
}
ASSERT((mcip->mci_flags &
MAC_CLIENT_FLAGS_PASSIVE_PRIMARY) == 0);
mcip->mci_flags |= MAC_CLIENT_FLAGS_PASSIVE_PRIMARY;
kmem_free(mrp, sizeof (*mrp));
/*
* Stash the unicast address handle, we will use it when
* we set up the passive client.
*/
mcip->mci_p_unicast_list = muip;
*mah = (mac_unicast_handle_t)muip;
return (0);
}
err = mac_client_datapath_setup(mcip, vid, mac_addr, mrp,
is_primary || is_vnic_primary, muip);
if (err != 0)
goto bail_out;
kmem_free(mrp, sizeof (*mrp));
*mah = (mac_unicast_handle_t)muip;
return (0);
bail_out:
if (fastpath_disabled)
mac_fastpath_enable((mac_handle_t)mip);
if (mcip->mci_state_flags & MCIS_EXCLUSIVE) {
mip->mi_state_flags &= ~MIS_EXCLUSIVE;
if (mip->mi_state_flags & MIS_LEGACY) {
mip->mi_capab_legacy.ml_active_clear(
mip->mi_driver);
}
}
kmem_free(mrp, sizeof (*mrp));
kmem_free(muip, sizeof (mac_unicast_impl_t));
return (err);
}
/*
* Wrapper function to mac_unicast_add when we want to have the same mac
* client open for two instances, one that is currently active and another
* that will become active when the current one is removed. In this case
* mac_unicast_add will return EGAIN and we will save the rx function and
* arg which will be used when we activate the passive client in
* mac_unicast_remove.
*/
int
mac_unicast_add_set_rx(mac_client_handle_t mch, uint8_t *mac_addr,
uint16_t flags, mac_unicast_handle_t *mah, uint16_t vid, mac_diag_t *diag,
mac_rx_t rx_fn, void *arg)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
uint_t err;
err = mac_unicast_add(mch, mac_addr, flags, mah, vid, diag);
if (err != 0 && err != EAGAIN)
return (err);
if (err == EAGAIN) {
if (rx_fn != NULL) {
mcip->mci_rx_p_fn = rx_fn;
mcip->mci_rx_p_arg = arg;
}
return (0);
}
if (rx_fn != NULL)
mac_rx_set(mch, rx_fn, arg);
return (err);
}
int
mac_unicast_add(mac_client_handle_t mch, uint8_t *mac_addr, uint16_t flags,
mac_unicast_handle_t *mah, uint16_t vid, mac_diag_t *diag)
{
mac_impl_t *mip = ((mac_client_impl_t *)mch)->mci_mip;
uint_t err;
i_mac_perim_enter(mip);
err = i_mac_unicast_add(mch, mac_addr, flags, mah, vid, diag);
i_mac_perim_exit(mip);
return (err);
}
static void
mac_client_datapath_teardown(mac_client_handle_t mch, mac_unicast_impl_t *muip,
flow_entry_t *flent)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
boolean_t no_unicast;
/*
* If we have not added a unicast address for this MAC client, just
* teardown the datapath.
*/
no_unicast = mcip->mci_state_flags & MCIS_NO_UNICAST_ADDR;
if (!no_unicast) {
/*
* We would have initialized subflows etc. only if we brought
* up the primary client and set the unicast unicast address
* etc. Deactivate the flows. The flow entry will be removed
* from the active flow tables, and the associated SRS,
* softrings etc will be deleted. But the flow entry itself
* won't be destroyed, instead it will continue to be archived
* off the the global flow hash list, for a possible future
* activation when say IP is plumbed again.
*/
mac_link_release_flows(mch);
}
mip->mi_nactiveclients--;
mac_update_single_active_client(mip);
/* Tear down the data path */
mac_datapath_teardown(mcip, mcip->mci_flent, SRST_LINK);
/*
* Prevent any future access to the flow entry through the mci_flent
* pointer by setting the mci_flent to NULL. Access to mci_flent in
* mac_bcast_send is also under mi_rw_lock.
*/
rw_enter(&mip->mi_rw_lock, RW_WRITER);
flent = mcip->mci_flent;
mac_client_remove_flow_from_list(mcip, flent);
if (mcip->mci_state_flags & MCIS_DESC_LOGGED)
mcip->mci_state_flags &= ~MCIS_DESC_LOGGED;
/*
* This is the last unicast address being removed and there shouldn't
* be any outbound data threads at this point coming down from mac
* clients. We have waited for the data threads to finish before
* starting dld_str_detach. Non-data threads must access TX SRS
* under mi_rw_lock.
*/
rw_exit(&mip->mi_rw_lock);
/*
* Don't use FLOW_MARK with FE_MC_NO_DATAPATH, as the flow might
* contain other flags, such as FE_CONDEMNED, which we need to
* cleared. We don't call mac_flow_cleanup() for this unicast
* flow as we have a already cleaned up SRSs etc. (via the teadown
* path). We just clear the stats and reset the initial callback
* function, the rest will be set when we call mac_flow_create,
* if at all.
*/
mutex_enter(&flent->fe_lock);
ASSERT(flent->fe_refcnt == 1 && flent->fe_mbg == NULL &&
flent->fe_tx_srs == NULL && flent->fe_rx_srs_cnt == 0);
flent->fe_flags = FE_MC_NO_DATAPATH;
flow_stat_destroy(flent);
mac_misc_stat_delete(flent);
/* Initialize the receiver function to a safe routine */
flent->fe_cb_fn = (flow_fn_t)mac_pkt_drop;
flent->fe_cb_arg1 = NULL;
flent->fe_cb_arg2 = NULL;
flent->fe_index = -1;
mutex_exit(&flent->fe_lock);
if (mip->mi_type->mt_brdcst_addr != NULL) {
ASSERT(muip != NULL || no_unicast);
mac_bcast_delete(mcip, mip->mi_type->mt_brdcst_addr,
muip != NULL ? muip->mui_vid : VLAN_ID_NONE);
}
if (mip->mi_nactiveclients == 1) {
mac_capab_update((mac_handle_t)mip);
mac_virtual_link_update(mip);
}
if (mcip->mci_state_flags & MCIS_EXCLUSIVE) {
mip->mi_state_flags &= ~MIS_EXCLUSIVE;
if (mip->mi_state_flags & MIS_LEGACY)
mip->mi_capab_legacy.ml_active_clear(mip->mi_driver);
}
mcip->mci_state_flags &= ~MCIS_UNICAST_HW;
if (mcip->mci_state_flags & MCIS_TAG_DISABLE)
mcip->mci_state_flags &= ~MCIS_TAG_DISABLE;
if (mcip->mci_state_flags & MCIS_STRIP_DISABLE)
mcip->mci_state_flags &= ~MCIS_STRIP_DISABLE;
if (mcip->mci_state_flags & MCIS_DISABLE_TX_VID_CHECK)
mcip->mci_state_flags &= ~MCIS_DISABLE_TX_VID_CHECK;
if (muip != NULL)
kmem_free(muip, sizeof (mac_unicast_impl_t));
mac_protect_cancel_timer(mcip);
mac_protect_flush_dynamic(mcip);
bzero(&mcip->mci_misc_stat, sizeof (mcip->mci_misc_stat));
/*
* Disable fastpath if this is a VNIC or a VLAN.
*/
if (mcip->mci_state_flags & MCIS_IS_VNIC)
mac_fastpath_enable((mac_handle_t)mip);
mac_stop((mac_handle_t)mip);
}
/*
* Remove a MAC address which was previously added by mac_unicast_add().
*/
int
mac_unicast_remove(mac_client_handle_t mch, mac_unicast_handle_t mah)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_unicast_impl_t *muip = (mac_unicast_impl_t *)mah;
mac_unicast_impl_t *pre;
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent;
uint16_t mui_vid;
i_mac_perim_enter(mip);
if (mcip->mci_flags & MAC_CLIENT_FLAGS_VNIC_PRIMARY) {
/*
* Called made by the upper MAC client of a VNIC.
* There's nothing much to do, the unicast address will
* be removed by the VNIC driver when the VNIC is deleted,
* but let's ensure that all our transmit is done before
* the client does a mac_client_stop lest it trigger an
* assert in the driver.
*/
ASSERT(muip->mui_vid == 0);
mac_tx_client_flush(mcip);
if ((mcip->mci_flags & MAC_CLIENT_FLAGS_PASSIVE_PRIMARY) != 0) {
mcip->mci_flags &= ~MAC_CLIENT_FLAGS_PASSIVE_PRIMARY;
if (mcip->mci_rx_p_fn != NULL) {
mac_rx_set(mch, mcip->mci_rx_p_fn,
mcip->mci_rx_p_arg);
mcip->mci_rx_p_fn = NULL;
mcip->mci_rx_p_arg = NULL;
}
kmem_free(muip, sizeof (mac_unicast_impl_t));
i_mac_perim_exit(mip);
return (0);
}
mcip->mci_flags &= ~MAC_CLIENT_FLAGS_VNIC_PRIMARY;
if (mcip->mci_state_flags & MCIS_TAG_DISABLE)
mcip->mci_state_flags &= ~MCIS_TAG_DISABLE;
if (mcip->mci_state_flags & MCIS_STRIP_DISABLE)
mcip->mci_state_flags &= ~MCIS_STRIP_DISABLE;
if (mcip->mci_state_flags & MCIS_DISABLE_TX_VID_CHECK)
mcip->mci_state_flags &= ~MCIS_DISABLE_TX_VID_CHECK;
kmem_free(muip, sizeof (mac_unicast_impl_t));
i_mac_perim_exit(mip);
return (0);
}
ASSERT(muip != NULL);
/*
* We are removing a passive client, we haven't setup the datapath
* for this yet, so nothing much to do.
*/
if ((mcip->mci_flags & MAC_CLIENT_FLAGS_PASSIVE_PRIMARY) != 0) {
ASSERT((mcip->mci_flent->fe_flags & FE_MC_NO_DATAPATH) != 0);
ASSERT(mcip->mci_p_unicast_list == muip);
mcip->mci_flags &= ~MAC_CLIENT_FLAGS_PASSIVE_PRIMARY;
mcip->mci_p_unicast_list = NULL;
mcip->mci_rx_p_fn = NULL;
mcip->mci_rx_p_arg = NULL;
mcip->mci_state_flags &= ~MCIS_UNICAST_HW;
if (mcip->mci_state_flags & MCIS_TAG_DISABLE)
mcip->mci_state_flags &= ~MCIS_TAG_DISABLE;
if (mcip->mci_state_flags & MCIS_STRIP_DISABLE)
mcip->mci_state_flags &= ~MCIS_STRIP_DISABLE;
if (mcip->mci_state_flags & MCIS_DISABLE_TX_VID_CHECK)
mcip->mci_state_flags &= ~MCIS_DISABLE_TX_VID_CHECK;
kmem_free(muip, sizeof (mac_unicast_impl_t));
i_mac_perim_exit(mip);
return (0);
}
/*
* Remove the VID from the list of client's VIDs.
*/
pre = mcip->mci_unicast_list;
if (muip == pre) {
mcip->mci_unicast_list = muip->mui_next;
} else {
while ((pre->mui_next != NULL) && (pre->mui_next != muip))
pre = pre->mui_next;
ASSERT(pre->mui_next == muip);
rw_enter(&mcip->mci_rw_lock, RW_WRITER);
pre->mui_next = muip->mui_next;
rw_exit(&mcip->mci_rw_lock);
}
if (!mac_client_single_rcvr(mcip)) {
/*
* This MAC client is shared by more than one unicast
* addresses, so we will just remove the flent
* corresponding to the address being removed. We don't invoke
* mac_rx_classify_flow_rem() since the additional flow is
* not associated with its own separate set of SRS and rings,
* and these constructs are still needed for the remaining
* flows.
*/
flent = mac_client_get_flow(mcip, muip);
ASSERT(flent != NULL);
/*
* The first one is disappearing, need to make sure
* we replace it with another from the list of
* shared clients.
*/
if (flent == mcip->mci_flent)
flent = mac_client_swap_mciflent(mcip);
mac_client_remove_flow_from_list(mcip, flent);
mac_flow_remove(mip->mi_flow_tab, flent, B_FALSE);
mac_flow_wait(flent, FLOW_DRIVER_UPCALL);
/*
* The multicast groups that were added by the client so
* far must be removed from the brodcast domain corresponding
* to the VID being removed.
*/
mac_client_bcast_refresh(mcip, mac_client_update_mcast,
(void *)flent, B_FALSE);
if (mip->mi_type->mt_brdcst_addr != NULL) {
mac_bcast_delete(mcip, mip->mi_type->mt_brdcst_addr,
muip->mui_vid);
}
FLOW_FINAL_REFRELE(flent);
ASSERT(!(mcip->mci_state_flags & MCIS_EXCLUSIVE));
/*
* Enable fastpath if this is a VNIC or a VLAN.
*/
if (mcip->mci_state_flags & MCIS_IS_VNIC)
mac_fastpath_enable((mac_handle_t)mip);
mac_stop((mac_handle_t)mip);
i_mac_perim_exit(mip);
return (0);
}
mui_vid = muip->mui_vid;
mac_client_datapath_teardown(mch, muip, flent);
if ((mcip->mci_flags & MAC_CLIENT_FLAGS_PRIMARY) && mui_vid == 0) {
mcip->mci_flags &= ~MAC_CLIENT_FLAGS_PRIMARY;
} else {
i_mac_perim_exit(mip);
return (0);
}
/*
* If we are removing the primary, check if we have a passive primary
* client that we need to activate now.
*/
mcip = mac_get_passive_primary_client(mip);
if (mcip != NULL) {
mac_resource_props_t *mrp;
mac_unicast_impl_t *muip;
mcip->mci_flags &= ~MAC_CLIENT_FLAGS_PASSIVE_PRIMARY;
mrp = kmem_zalloc(sizeof (*mrp), KM_SLEEP);
/*
* Apply the property cached in the mac_impl_t to the
* primary mac client.
*/
mac_get_resources((mac_handle_t)mip, mrp);
(void) mac_client_set_resources(mch, mrp);
ASSERT(mcip->mci_p_unicast_list != NULL);
muip = mcip->mci_p_unicast_list;
mcip->mci_p_unicast_list = NULL;
if (mac_client_datapath_setup(mcip, VLAN_ID_NONE,
mip->mi_addr, mrp, B_TRUE, muip) == 0) {
if (mcip->mci_rx_p_fn != NULL) {
mac_rx_set(mch, mcip->mci_rx_p_fn,
mcip->mci_rx_p_arg);
mcip->mci_rx_p_fn = NULL;
mcip->mci_rx_p_arg = NULL;
}
} else {
kmem_free(muip, sizeof (mac_unicast_impl_t));
}
kmem_free(mrp, sizeof (*mrp));
}
i_mac_perim_exit(mip);
return (0);
}
/*
* Multicast add function invoked by MAC clients.
*/
int
mac_multicast_add(mac_client_handle_t mch, const uint8_t *addr)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent = mcip->mci_flent_list;
flow_entry_t *prev_fe = NULL;
uint16_t vid;
int err = 0;
/* Verify the address is a valid multicast address */
if ((err = mip->mi_type->mt_ops.mtops_multicst_verify(addr,
mip->mi_pdata)) != 0)
return (err);
i_mac_perim_enter(mip);
while (flent != NULL) {
vid = i_mac_flow_vid(flent);
err = mac_bcast_add((mac_client_impl_t *)mch, addr, vid,
MAC_ADDRTYPE_MULTICAST);
if (err != 0)
break;
prev_fe = flent;
flent = flent->fe_client_next;
}
/*
* If we failed adding, then undo all, rather than partial
* success.
*/
if (flent != NULL && prev_fe != NULL) {
flent = mcip->mci_flent_list;
while (flent != prev_fe->fe_client_next) {
vid = i_mac_flow_vid(flent);
mac_bcast_delete((mac_client_impl_t *)mch, addr, vid);
flent = flent->fe_client_next;
}
}
i_mac_perim_exit(mip);
return (err);
}
/*
* Multicast delete function invoked by MAC clients.
*/
void
mac_multicast_remove(mac_client_handle_t mch, const uint8_t *addr)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
flow_entry_t *flent;
uint16_t vid;
i_mac_perim_enter(mip);
for (flent = mcip->mci_flent_list; flent != NULL;
flent = flent->fe_client_next) {
vid = i_mac_flow_vid(flent);
mac_bcast_delete((mac_client_impl_t *)mch, addr, vid);
}
i_mac_perim_exit(mip);
}
/*
* When a MAC client desires to capture packets on an interface,
* it registers a promiscuous call back with mac_promisc_add().
* There are three types of promiscuous callbacks:
*
* * MAC_CLIENT_PROMISC_ALL
* Captures all packets sent and received by the MAC client,
* the physical interface, as well as all other MAC clients
* defined on top of the same MAC.
*
* * MAC_CLIENT_PROMISC_FILTERED
* Captures all packets sent and received by the MAC client,
* plus all multicast traffic sent and received by the phyisical
* interface and the other MAC clients.
*
* * MAC_CLIENT_PROMISC_MULTI
* Captures all broadcast and multicast packets sent and
* received by the MAC clients as well as the physical interface.
*
* In all cases, the underlying MAC is put in promiscuous mode.
*/
int
mac_promisc_add(mac_client_handle_t mch, mac_client_promisc_type_t type,
mac_rx_t fn, void *arg, mac_promisc_handle_t *mphp, uint16_t flags)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
mac_promisc_impl_t *mpip;
mac_cb_info_t *mcbi;
int rc;
i_mac_perim_enter(mip);
if ((rc = mac_start((mac_handle_t)mip)) != 0) {
i_mac_perim_exit(mip);
return (rc);
}
if ((mcip->mci_state_flags & MCIS_IS_VNIC) &&
type == MAC_CLIENT_PROMISC_ALL &&
(mcip->mci_protect_flags & MPT_FLAG_PROMISC_FILTERED)) {
/*
* The function is being invoked by the upper MAC client
* of a VNIC. The VNIC should only see the traffic
* it is entitled to.
*/
type = MAC_CLIENT_PROMISC_FILTERED;
}
/*
* Turn on promiscuous mode for the underlying NIC.
* This is needed even for filtered callbacks which
* expect to receive all multicast traffic on the wire.
*
* Physical promiscuous mode should not be turned on if
* MAC_PROMISC_FLAGS_NO_PHYS is set.
*/
if ((flags & MAC_PROMISC_FLAGS_NO_PHYS) == 0) {
if ((rc = i_mac_promisc_set(mip, B_TRUE)) != 0) {
mac_stop((mac_handle_t)mip);
i_mac_perim_exit(mip);
return (rc);
}
}
mpip = kmem_cache_alloc(mac_promisc_impl_cache, KM_SLEEP);
mpip->mpi_type = type;
mpip->mpi_fn = fn;
mpip->mpi_arg = arg;
mpip->mpi_mcip = mcip;
mpip->mpi_no_tx_loop = ((flags & MAC_PROMISC_FLAGS_NO_TX_LOOP) != 0);
mpip->mpi_no_phys = ((flags & MAC_PROMISC_FLAGS_NO_PHYS) != 0);
mpip->mpi_strip_vlan_tag =
((flags & MAC_PROMISC_FLAGS_VLAN_TAG_STRIP) != 0);
mpip->mpi_no_copy = ((flags & MAC_PROMISC_FLAGS_NO_COPY) != 0);
mcbi = &mip->mi_promisc_cb_info;
mutex_enter(mcbi->mcbi_lockp);
mac_callback_add(&mip->mi_promisc_cb_info, &mcip->mci_promisc_list,
&mpip->mpi_mci_link);
mac_callback_add(&mip->mi_promisc_cb_info, &mip->mi_promisc_list,
&mpip->mpi_mi_link);
mutex_exit(mcbi->mcbi_lockp);
*mphp = (mac_promisc_handle_t)mpip;
if (mcip->mci_state_flags & MCIS_IS_VNIC) {
mac_impl_t *umip = mcip->mci_upper_mip;
ASSERT(umip != NULL);
mac_vnic_secondary_update(umip);
}
i_mac_perim_exit(mip);
return (0);
}
/*
* Remove a multicast address previously aded through mac_promisc_add().
*/
void
mac_promisc_remove(mac_promisc_handle_t mph)
{
mac_promisc_impl_t *mpip = (mac_promisc_impl_t *)mph;
mac_client_impl_t *mcip = mpip->mpi_mcip;
mac_impl_t *mip = mcip->mci_mip;
mac_cb_info_t *mcbi;
int rv;
i_mac_perim_enter(mip);
/*
* Even if the device can't be reset into normal mode, we still
* need to clear the client promisc callbacks. The client may want
* to close the mac end point and we can't have stale callbacks.
*/
if (!(mpip->mpi_no_phys)) {
if ((rv = i_mac_promisc_set(mip, B_FALSE)) != 0) {
cmn_err(CE_WARN, "%s: failed to switch OFF promiscuous"
" mode because of error 0x%x", mip->mi_name, rv);
}
}
mcbi = &mip->mi_promisc_cb_info;
mutex_enter(mcbi->mcbi_lockp);
if (mac_callback_remove(mcbi, &mip->mi_promisc_list,
&mpip->mpi_mi_link)) {
VERIFY(mac_callback_remove(&mip->mi_promisc_cb_info,
&mcip->mci_promisc_list, &mpip->mpi_mci_link));
kmem_cache_free(mac_promisc_impl_cache, mpip);
} else {
mac_callback_remove_wait(&mip->mi_promisc_cb_info);
}
if (mcip->mci_state_flags & MCIS_IS_VNIC) {
mac_impl_t *umip = mcip->mci_upper_mip;
ASSERT(umip != NULL);
mac_vnic_secondary_update(umip);
}
mutex_exit(mcbi->mcbi_lockp);
mac_stop((mac_handle_t)mip);
i_mac_perim_exit(mip);
}
/*
* Reference count the number of active Tx threads. MCI_TX_QUIESCE indicates
* that a control operation wants to quiesce the Tx data flow in which case
* we return an error. Holding any of the per cpu locks ensures that the
* mci_tx_flag won't change.
*
* 'CPU' must be accessed just once and used to compute the index into the
* percpu array, and that index must be used for the entire duration of the
* packet send operation. Note that the thread may be preempted and run on
* another cpu any time and so we can't use 'CPU' more than once for the
* operation.
*/
#define MAC_TX_TRY_HOLD(mcip, mytx, error) \
{ \
(error) = 0; \
(mytx) = &(mcip)->mci_tx_pcpu[CPU->cpu_seqid & mac_tx_percpu_cnt]; \
mutex_enter(&(mytx)->pcpu_tx_lock); \
if (!((mcip)->mci_tx_flag & MCI_TX_QUIESCE)) { \
(mytx)->pcpu_tx_refcnt++; \
} else { \
(error) = -1; \
} \
mutex_exit(&(mytx)->pcpu_tx_lock); \
}
/*
* Release the reference. If needed, signal any control operation waiting
* for Tx quiescence. The wait and signal are always done using the
* mci_tx_pcpu[0]'s lock
*/
#define MAC_TX_RELE(mcip, mytx) { \
mutex_enter(&(mytx)->pcpu_tx_lock); \
if (--(mytx)->pcpu_tx_refcnt == 0 && \
(mcip)->mci_tx_flag & MCI_TX_QUIESCE) { \
mutex_exit(&(mytx)->pcpu_tx_lock); \
mutex_enter(&(mcip)->mci_tx_pcpu[0].pcpu_tx_lock); \
cv_signal(&(mcip)->mci_tx_cv); \
mutex_exit(&(mcip)->mci_tx_pcpu[0].pcpu_tx_lock); \
} else { \
mutex_exit(&(mytx)->pcpu_tx_lock); \
} \
}
/*
* Send function invoked by MAC clients.
*/
mac_tx_cookie_t
mac_tx(mac_client_handle_t mch, mblk_t *mp_chain, uintptr_t hint,
uint16_t flag, mblk_t **ret_mp)
{
mac_tx_cookie_t cookie = NULL;
int error;
mac_tx_percpu_t *mytx;
mac_soft_ring_set_t *srs;
flow_entry_t *flent;
boolean_t is_subflow = B_FALSE;
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
mac_srs_tx_t *srs_tx;
/*
* Check whether the active Tx threads count is bumped already.
*/
if (!(flag & MAC_TX_NO_HOLD)) {
MAC_TX_TRY_HOLD(mcip, mytx, error);
if (error != 0) {
freemsgchain(mp_chain);
return (NULL);
}
}
/*
* If mac protection is enabled, only the permissible packets will be
* returned by mac_protect_check().
*/
if ((mcip->mci_flent->
fe_resource_props.mrp_mask & MRP_PROTECT) != 0 &&
(mp_chain = mac_protect_check(mch, mp_chain)) == NULL)
goto done;
if (mcip->mci_subflow_tab != NULL &&
mcip->mci_subflow_tab->ft_flow_count > 0 &&
mac_flow_lookup(mcip->mci_subflow_tab, mp_chain,
FLOW_OUTBOUND, &flent) == 0) {
/*
* The main assumption here is that if in the event
* we get a chain, all the packets will be classified
* to the same Flow/SRS. If this changes for any
* reason, the following logic should change as well.
* I suppose the fanout_hint also assumes this .
*/
ASSERT(flent != NULL);
is_subflow = B_TRUE;
} else {
flent = mcip->mci_flent;
}
srs = flent->fe_tx_srs;
/*
* This is to avoid panics with PF_PACKET that can call mac_tx()
* against an interface that is not capable of sending. A rewrite
* of the mac datapath is required to remove this limitation.
*/
if (srs == NULL) {
freemsgchain(mp_chain);
goto done;
}
srs_tx = &srs->srs_tx;
if (srs_tx->st_mode == SRS_TX_DEFAULT &&
(srs->srs_state & SRS_ENQUEUED) == 0 &&
mip->mi_nactiveclients == 1 && mp_chain->b_next == NULL) {
uint64_t obytes;
/*
* Since dls always opens the underlying MAC, nclients equals
* to 1 means that the only active client is dls itself acting
* as a primary client of the MAC instance. Since dls will not
* send tagged packets in that case, and dls is trusted to send
* packets for its allowed VLAN(s), the VLAN tag insertion and
* check is required only if nclients is greater than 1.
*/
if (mip->mi_nclients > 1) {
if (MAC_VID_CHECK_NEEDED(mcip)) {
int err = 0;
MAC_VID_CHECK(mcip, mp_chain, err);
if (err != 0) {
freemsg(mp_chain);
mcip->mci_misc_stat.mms_txerrors++;
goto done;
}
}
if (MAC_TAG_NEEDED(mcip)) {
mp_chain = mac_add_vlan_tag(mp_chain, 0,
mac_client_vid(mch));
if (mp_chain == NULL) {
mcip->mci_misc_stat.mms_txerrors++;
goto done;
}
}
}
obytes = (mp_chain->b_cont == NULL ? MBLKL(mp_chain) :
msgdsize(mp_chain));
MAC_TX(mip, srs_tx->st_arg2, mp_chain, mcip);
if (mp_chain == NULL) {
cookie = NULL;
SRS_TX_STAT_UPDATE(srs, opackets, 1);
SRS_TX_STAT_UPDATE(srs, obytes, obytes);
} else {
mutex_enter(&srs->srs_lock);
cookie = mac_tx_srs_no_desc(srs, mp_chain,
flag, ret_mp);
mutex_exit(&srs->srs_lock);
}
} else {
cookie = srs_tx->st_func(srs, mp_chain, hint, flag, ret_mp);
}
done:
if (is_subflow)
FLOW_REFRELE(flent);
if (!(flag & MAC_TX_NO_HOLD))
MAC_TX_RELE(mcip, mytx);
return (cookie);
}
/*
* mac_tx_is_blocked
*
* Given a cookie, it returns if the ring identified by the cookie is
* flow-controlled or not. If NULL is passed in place of a cookie,
* then it finds out if any of the underlying rings belonging to the
* SRS is flow controlled or not and returns that status.
*/
/* ARGSUSED */
boolean_t
mac_tx_is_flow_blocked(mac_client_handle_t mch, mac_tx_cookie_t cookie)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_soft_ring_set_t *mac_srs;
mac_soft_ring_t *sringp;
boolean_t blocked = B_FALSE;
mac_tx_percpu_t *mytx;
int err;
int i;
/*
* Bump the reference count so that mac_srs won't be deleted.
* If the client is currently quiesced and we failed to bump
* the reference, return B_TRUE so that flow control stays
* as enabled.
*
* Flow control will then be disabled once the client is no
* longer quiesced.
*/
MAC_TX_TRY_HOLD(mcip, mytx, err);
if (err != 0)
return (B_TRUE);
if ((mac_srs = MCIP_TX_SRS(mcip)) == NULL) {
MAC_TX_RELE(mcip, mytx);
return (B_FALSE);
}
mutex_enter(&mac_srs->srs_lock);
/*
* Only in the case of TX_FANOUT and TX_AGGR, the underlying
* softring (s_ring_state) will have the HIWAT set. This is
* the multiple Tx ring flow control case. For all other
* case, SRS (srs_state) will store the condition.
*/
if (mac_srs->srs_tx.st_mode == SRS_TX_FANOUT ||
mac_srs->srs_tx.st_mode == SRS_TX_AGGR) {
if (cookie != NULL) {
sringp = (mac_soft_ring_t *)cookie;
mutex_enter(&sringp->s_ring_lock);
if (sringp->s_ring_state & S_RING_TX_HIWAT)
blocked = B_TRUE;
mutex_exit(&sringp->s_ring_lock);
} else {
for (i = 0; i < mac_srs->srs_tx_ring_count; i++) {
sringp = mac_srs->srs_tx_soft_rings[i];
mutex_enter(&sringp->s_ring_lock);
if (sringp->s_ring_state & S_RING_TX_HIWAT) {
blocked = B_TRUE;
mutex_exit(&sringp->s_ring_lock);
break;
}
mutex_exit(&sringp->s_ring_lock);
}
}
} else {
blocked = (mac_srs->srs_state & SRS_TX_HIWAT);
}
mutex_exit(&mac_srs->srs_lock);
MAC_TX_RELE(mcip, mytx);
return (blocked);
}
/*
* Check if the MAC client is the primary MAC client.
*/
boolean_t
mac_is_primary_client(mac_client_impl_t *mcip)
{
return (mcip->mci_flags & MAC_CLIENT_FLAGS_PRIMARY);
}
void
mac_ioctl(mac_handle_t mh, queue_t *wq, mblk_t *bp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
int cmd = ((struct iocblk *)bp->b_rptr)->ioc_cmd;
if ((cmd == ND_GET && (mip->mi_callbacks->mc_callbacks & MC_GETPROP)) ||
(cmd == ND_SET && (mip->mi_callbacks->mc_callbacks & MC_SETPROP))) {
/*
* If ndd props were registered, call them.
* Note that ndd ioctls are Obsolete
*/
mac_ndd_ioctl(mip, wq, bp);
return;
}
/*
* Call the driver to handle the ioctl. The driver may not support
* any ioctls, in which case we reply with a NAK on its behalf.
*/
if (mip->mi_callbacks->mc_callbacks & MC_IOCTL)
mip->mi_ioctl(mip->mi_driver, wq, bp);
else
miocnak(wq, bp, 0, EINVAL);
}
/*
* Return the link state of the specified MAC instance.
*/
link_state_t
mac_link_get(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_linkstate);
}
/*
* Add a mac client specified notification callback. Please see the comments
* above mac_callback_add() for general information about mac callback
* addition/deletion in the presence of mac callback list walkers
*/
mac_notify_handle_t
mac_notify_add(mac_handle_t mh, mac_notify_t notify_fn, void *arg)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_notify_cb_t *mncb;
mac_cb_info_t *mcbi;
/*
* Allocate a notify callback structure, fill in the details and
* use the mac callback list manipulation functions to chain into
* the list of callbacks.
*/
mncb = kmem_zalloc(sizeof (mac_notify_cb_t), KM_SLEEP);
mncb->mncb_fn = notify_fn;
mncb->mncb_arg = arg;
mncb->mncb_mip = mip;
mncb->mncb_link.mcb_objp = mncb;
mncb->mncb_link.mcb_objsize = sizeof (mac_notify_cb_t);
mncb->mncb_link.mcb_flags = MCB_NOTIFY_CB_T;
mcbi = &mip->mi_notify_cb_info;
i_mac_perim_enter(mip);
mutex_enter(mcbi->mcbi_lockp);
mac_callback_add(&mip->mi_notify_cb_info, &mip->mi_notify_cb_list,
&mncb->mncb_link);
mutex_exit(mcbi->mcbi_lockp);
i_mac_perim_exit(mip);
return ((mac_notify_handle_t)mncb);
}
void
mac_notify_remove_wait(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_cb_info_t *mcbi = &mip->mi_notify_cb_info;
mutex_enter(mcbi->mcbi_lockp);
mac_callback_remove_wait(&mip->mi_notify_cb_info);
mutex_exit(mcbi->mcbi_lockp);
}
/*
* Remove a mac client specified notification callback
*/
int
mac_notify_remove(mac_notify_handle_t mnh, boolean_t wait)
{
mac_notify_cb_t *mncb = (mac_notify_cb_t *)mnh;
mac_impl_t *mip = mncb->mncb_mip;
mac_cb_info_t *mcbi;
int err = 0;
mcbi = &mip->mi_notify_cb_info;
i_mac_perim_enter(mip);
mutex_enter(mcbi->mcbi_lockp);
ASSERT(mncb->mncb_link.mcb_objp == mncb);
/*
* If there aren't any list walkers, the remove would succeed
* inline, else we wait for the deferred remove to complete
*/
if (mac_callback_remove(&mip->mi_notify_cb_info,
&mip->mi_notify_cb_list, &mncb->mncb_link)) {
kmem_free(mncb, sizeof (mac_notify_cb_t));
} else {
err = EBUSY;
}
mutex_exit(mcbi->mcbi_lockp);
i_mac_perim_exit(mip);
/*
* If we failed to remove the notification callback and "wait" is set
* to be B_TRUE, wait for the callback to finish after we exit the
* mac perimeter.
*/
if (err != 0 && wait) {
mac_notify_remove_wait((mac_handle_t)mip);
return (0);
}
return (err);
}
/*
* Associate resource management callbacks with the specified MAC
* clients.
*/
void
mac_resource_set_common(mac_client_handle_t mch, mac_resource_add_t add,
mac_resource_remove_t remove, mac_resource_quiesce_t quiesce,
mac_resource_restart_t restart, mac_resource_bind_t bind,
void *arg)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mcip->mci_resource_add = add;
mcip->mci_resource_remove = remove;
mcip->mci_resource_quiesce = quiesce;
mcip->mci_resource_restart = restart;
mcip->mci_resource_bind = bind;
mcip->mci_resource_arg = arg;
}
void
mac_resource_set(mac_client_handle_t mch, mac_resource_add_t add, void *arg)
{
/* update the 'resource_add' callback */
mac_resource_set_common(mch, add, NULL, NULL, NULL, NULL, arg);
}
/*
* Sets up the client resources and enable the polling interface over all the
* SRS's and the soft rings of the client
*/
void
mac_client_poll_enable(mac_client_handle_t mch)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_soft_ring_set_t *mac_srs;
flow_entry_t *flent;
int i;
flent = mcip->mci_flent;
ASSERT(flent != NULL);
mcip->mci_state_flags |= MCIS_CLIENT_POLL_CAPABLE;
for (i = 0; i < flent->fe_rx_srs_cnt; i++) {
mac_srs = (mac_soft_ring_set_t *)flent->fe_rx_srs[i];
ASSERT(mac_srs->srs_mcip == mcip);
mac_srs_client_poll_enable(mcip, mac_srs);
}
}
/*
* Tears down the client resources and disable the polling interface over all
* the SRS's and the soft rings of the client
*/
void
mac_client_poll_disable(mac_client_handle_t mch)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_soft_ring_set_t *mac_srs;
flow_entry_t *flent;
int i;
flent = mcip->mci_flent;
ASSERT(flent != NULL);
mcip->mci_state_flags &= ~MCIS_CLIENT_POLL_CAPABLE;
for (i = 0; i < flent->fe_rx_srs_cnt; i++) {
mac_srs = (mac_soft_ring_set_t *)flent->fe_rx_srs[i];
ASSERT(mac_srs->srs_mcip == mcip);
mac_srs_client_poll_disable(mcip, mac_srs);
}
}
/*
* Associate the CPUs specified by the given property with a MAC client.
*/
int
mac_cpu_set(mac_client_handle_t mch, mac_resource_props_t *mrp)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
int err = 0;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
if ((err = mac_validate_props(mcip->mci_state_flags & MCIS_IS_VNIC ?
mcip->mci_upper_mip : mip, mrp)) != 0) {
return (err);
}
if (MCIP_DATAPATH_SETUP(mcip))
mac_flow_modify(mip->mi_flow_tab, mcip->mci_flent, mrp);
mac_update_resources(mrp, MCIP_RESOURCE_PROPS(mcip), B_FALSE);
return (0);
}
/*
* Apply the specified properties to the specified MAC client.
*/
int
mac_client_set_resources(mac_client_handle_t mch, mac_resource_props_t *mrp)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = mcip->mci_mip;
int err = 0;
i_mac_perim_enter(mip);
if ((mrp->mrp_mask & MRP_MAXBW) || (mrp->mrp_mask & MRP_PRIORITY)) {
err = mac_resource_ctl_set(mch, mrp);
if (err != 0)
goto done;
}
if (mrp->mrp_mask & (MRP_CPUS|MRP_POOL)) {
err = mac_cpu_set(mch, mrp);
if (err != 0)
goto done;
}
if (mrp->mrp_mask & MRP_PROTECT) {
err = mac_protect_set(mch, mrp);
if (err != 0)
goto done;
}
if ((mrp->mrp_mask & MRP_RX_RINGS) || (mrp->mrp_mask & MRP_TX_RINGS))
err = mac_resource_ctl_set(mch, mrp);
done:
i_mac_perim_exit(mip);
return (err);
}
/*
* Return the properties currently associated with the specified MAC client.
*/
void
mac_client_get_resources(mac_client_handle_t mch, mac_resource_props_t *mrp)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_resource_props_t *mcip_mrp = MCIP_RESOURCE_PROPS(mcip);
bcopy(mcip_mrp, mrp, sizeof (mac_resource_props_t));
}
/*
* Return the effective properties currently associated with the specified
* MAC client.
*/
void
mac_client_get_effective_resources(mac_client_handle_t mch,
mac_resource_props_t *mrp)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_resource_props_t *mcip_mrp = MCIP_EFFECTIVE_PROPS(mcip);
bcopy(mcip_mrp, mrp, sizeof (mac_resource_props_t));
}
/*
* Pass a copy of the specified packet to the promiscuous callbacks
* of the specified MAC.
*
* If sender is NULL, the function is being invoked for a packet chain
* received from the wire. If sender is non-NULL, it points to
* the MAC client from which the packet is being sent.
*
* The packets are distributed to the promiscuous callbacks as follows:
*
* - all packets are sent to the MAC_CLIENT_PROMISC_ALL callbacks
* - all broadcast and multicast packets are sent to the
* MAC_CLIENT_PROMISC_FILTER and MAC_CLIENT_PROMISC_MULTI.
*
* The unicast packets of MAC_CLIENT_PROMISC_FILTER callbacks are dispatched
* after classification by mac_rx_deliver().
*/
static void
mac_promisc_dispatch_one(mac_promisc_impl_t *mpip, mblk_t *mp,
boolean_t loopback)
{
mblk_t *mp_copy, *mp_next;
if (!mpip->mpi_no_copy || mpip->mpi_strip_vlan_tag) {
mp_copy = copymsg(mp);
if (mp_copy == NULL)
return;
if (mpip->mpi_strip_vlan_tag) {
mp_copy = mac_strip_vlan_tag_chain(mp_copy);
if (mp_copy == NULL)
return;
}
mp_next = NULL;
} else {
mp_copy = mp;
mp_next = mp->b_next;
}
mp_copy->b_next = NULL;
mpip->mpi_fn(mpip->mpi_arg, NULL, mp_copy, loopback);
if (mp_copy == mp)
mp->b_next = mp_next;
}
/*
* Return the VID of a packet. Zero if the packet is not tagged.
*/
static uint16_t
mac_ether_vid(mblk_t *mp)
{
struct ether_header *eth = (struct ether_header *)mp->b_rptr;
if (ntohs(eth->ether_type) == ETHERTYPE_VLAN) {
struct ether_vlan_header *t_evhp =
(struct ether_vlan_header *)mp->b_rptr;
return (VLAN_ID(ntohs(t_evhp->ether_tci)));
}
return (0);
}
/*
* Return whether the specified packet contains a multicast or broadcast
* destination MAC address.
*/
static boolean_t
mac_is_mcast(mac_impl_t *mip, mblk_t *mp)
{
mac_header_info_t hdr_info;
if (mac_header_info((mac_handle_t)mip, mp, &hdr_info) != 0)
return (B_FALSE);
return ((hdr_info.mhi_dsttype == MAC_ADDRTYPE_BROADCAST) ||
(hdr_info.mhi_dsttype == MAC_ADDRTYPE_MULTICAST));
}
/*
* Send a copy of an mblk chain to the MAC clients of the specified MAC.
* "sender" points to the sender MAC client for outbound packets, and
* is set to NULL for inbound packets.
*/
void
mac_promisc_dispatch(mac_impl_t *mip, mblk_t *mp_chain,
mac_client_impl_t *sender)
{
mac_promisc_impl_t *mpip;
mac_cb_t *mcb;
mblk_t *mp;
boolean_t is_mcast, is_sender;
MAC_PROMISC_WALKER_INC(mip);
for (mp = mp_chain; mp != NULL; mp = mp->b_next) {
is_mcast = mac_is_mcast(mip, mp);
/* send packet to interested callbacks */
for (mcb = mip->mi_promisc_list; mcb != NULL;
mcb = mcb->mcb_nextp) {
mpip = (mac_promisc_impl_t *)mcb->mcb_objp;
is_sender = (mpip->mpi_mcip == sender);
if (is_sender && mpip->mpi_no_tx_loop)
/*
* The sender doesn't want to receive
* copies of the packets it sends.
*/
continue;
/* this client doesn't need any packets (bridge) */
if (mpip->mpi_fn == NULL)
continue;
/*
* For an ethernet MAC, don't displatch a multicast
* packet to a non-PROMISC_ALL callbacks unless the VID
* of the packet matches the VID of the client.
*/
if (is_mcast &&
mpip->mpi_type != MAC_CLIENT_PROMISC_ALL &&
!mac_client_check_flow_vid(mpip->mpi_mcip,
mac_ether_vid(mp)))
continue;
if (is_sender ||
mpip->mpi_type == MAC_CLIENT_PROMISC_ALL ||
is_mcast)
mac_promisc_dispatch_one(mpip, mp, is_sender);
}
}
MAC_PROMISC_WALKER_DCR(mip);
}
void
mac_promisc_client_dispatch(mac_client_impl_t *mcip, mblk_t *mp_chain)
{
mac_impl_t *mip = mcip->mci_mip;
mac_promisc_impl_t *mpip;
boolean_t is_mcast;
mblk_t *mp;
mac_cb_t *mcb;
/*
* The unicast packets for the MAC client still
* need to be delivered to the MAC_CLIENT_PROMISC_FILTERED
* promiscuous callbacks. The broadcast and multicast
* packets were delivered from mac_rx().
*/
MAC_PROMISC_WALKER_INC(mip);
for (mp = mp_chain; mp != NULL; mp = mp->b_next) {
is_mcast = mac_is_mcast(mip, mp);
for (mcb = mcip->mci_promisc_list; mcb != NULL;
mcb = mcb->mcb_nextp) {
mpip = (mac_promisc_impl_t *)mcb->mcb_objp;
if (mpip->mpi_type == MAC_CLIENT_PROMISC_FILTERED &&
!is_mcast) {
mac_promisc_dispatch_one(mpip, mp, B_FALSE);
}
}
}
MAC_PROMISC_WALKER_DCR(mip);
}
/*
* Return the margin value currently assigned to the specified MAC instance.
*/
void
mac_margin_get(mac_handle_t mh, uint32_t *marginp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
rw_enter(&(mip->mi_rw_lock), RW_READER);
*marginp = mip->mi_margin;
rw_exit(&(mip->mi_rw_lock));
}
/*
* mac_info_get() is used for retrieving the mac_info when a DL_INFO_REQ is
* issued before a DL_ATTACH_REQ. we walk the i_mac_impl_hash table and find
* the first mac_impl_t with a matching driver name; then we copy its mac_info_t
* to the caller. we do all this with i_mac_impl_lock held so the mac_impl_t
* cannot disappear while we are accessing it.
*/
typedef struct i_mac_info_state_s {
const char *mi_name;
mac_info_t *mi_infop;
} i_mac_info_state_t;
/*ARGSUSED*/
static uint_t
i_mac_info_walker(mod_hash_key_t key, mod_hash_val_t *val, void *arg)
{
i_mac_info_state_t *statep = arg;
mac_impl_t *mip = (mac_impl_t *)val;
if (mip->mi_state_flags & MIS_DISABLED)
return (MH_WALK_CONTINUE);
if (strcmp(statep->mi_name,
ddi_driver_name(mip->mi_dip)) != 0)
return (MH_WALK_CONTINUE);
statep->mi_infop = &mip->mi_info;
return (MH_WALK_TERMINATE);
}
boolean_t
mac_info_get(const char *name, mac_info_t *minfop)
{
i_mac_info_state_t state;
rw_enter(&i_mac_impl_lock, RW_READER);
state.mi_name = name;
state.mi_infop = NULL;
mod_hash_walk(i_mac_impl_hash, i_mac_info_walker, &state);
if (state.mi_infop == NULL) {
rw_exit(&i_mac_impl_lock);
return (B_FALSE);
}
*minfop = *state.mi_infop;
rw_exit(&i_mac_impl_lock);
return (B_TRUE);
}
/*
* To get the capabilities that MAC layer cares about, such as rings, factory
* mac address, vnic or not, it should directly invoke this function. If the
* link is part of a bridge, then the only "capability" it has is the inability
* to do zero copy.
*/
boolean_t
i_mac_capab_get(mac_handle_t mh, mac_capab_t cap, void *cap_data)
{
mac_impl_t *mip = (mac_impl_t *)mh;
if (mip->mi_bridge_link != NULL)
return (cap == MAC_CAPAB_NO_ZCOPY);
else if (mip->mi_callbacks->mc_callbacks & MC_GETCAPAB)
return (mip->mi_getcapab(mip->mi_driver, cap, cap_data));
else
return (B_FALSE);
}
/*
* Capability query function. If number of active mac clients is greater than
* 1, only limited capabilities can be advertised to the caller no matter the
* driver has certain capability or not. Else, we query the driver to get the
* capability.
*/
boolean_t
mac_capab_get(mac_handle_t mh, mac_capab_t cap, void *cap_data)
{
mac_impl_t *mip = (mac_impl_t *)mh;
/*
* if mi_nactiveclients > 1, only MAC_CAPAB_LEGACY, MAC_CAPAB_HCKSUM,
* MAC_CAPAB_NO_NATIVEVLAN and MAC_CAPAB_NO_ZCOPY can be advertised.
*/
if (mip->mi_nactiveclients > 1) {
switch (cap) {
case MAC_CAPAB_NO_ZCOPY:
return (B_TRUE);
case MAC_CAPAB_LEGACY:
case MAC_CAPAB_HCKSUM:
case MAC_CAPAB_NO_NATIVEVLAN:
break;
default:
return (B_FALSE);
}
}
/* else get capab from driver */
return (i_mac_capab_get(mh, cap, cap_data));
}
boolean_t
mac_sap_verify(mac_handle_t mh, uint32_t sap, uint32_t *bind_sap)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_type->mt_ops.mtops_sap_verify(sap, bind_sap,
mip->mi_pdata));
}
mblk_t *
mac_header(mac_handle_t mh, const uint8_t *daddr, uint32_t sap, mblk_t *payload,
size_t extra_len)
{
mac_impl_t *mip = (mac_impl_t *)mh;
const uint8_t *hdr_daddr;
/*
* If the MAC is point-to-point with a fixed destination address, then
* we must always use that destination in the MAC header.
*/
hdr_daddr = (mip->mi_dstaddr_set ? mip->mi_dstaddr : daddr);
return (mip->mi_type->mt_ops.mtops_header(mip->mi_addr, hdr_daddr, sap,
mip->mi_pdata, payload, extra_len));
}
int
mac_header_info(mac_handle_t mh, mblk_t *mp, mac_header_info_t *mhip)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_type->mt_ops.mtops_header_info(mp, mip->mi_pdata,
mhip));
}
int
mac_vlan_header_info(mac_handle_t mh, mblk_t *mp, mac_header_info_t *mhip)
{
mac_impl_t *mip = (mac_impl_t *)mh;
boolean_t is_ethernet = (mip->mi_info.mi_media == DL_ETHER);
int err = 0;
/*
* Packets should always be at least 16 bit aligned.
*/
ASSERT(IS_P2ALIGNED(mp->b_rptr, sizeof (uint16_t)));
if ((err = mac_header_info(mh, mp, mhip)) != 0)
return (err);
/*
* If this is a VLAN-tagged Ethernet packet, then the SAP in the
* mac_header_info_t as returned by mac_header_info() is
* ETHERTYPE_VLAN. We need to grab the ethertype from the VLAN header.
*/
if (is_ethernet && (mhip->mhi_bindsap == ETHERTYPE_VLAN)) {
struct ether_vlan_header *evhp;
uint16_t sap;
mblk_t *tmp = NULL;
size_t size;
size = sizeof (struct ether_vlan_header);
if (MBLKL(mp) < size) {
/*
* Pullup the message in order to get the MAC header
* infomation. Note that this is a read-only function,
* we keep the input packet intact.
*/
if ((tmp = msgpullup(mp, size)) == NULL)
return (EINVAL);
mp = tmp;
}
evhp = (struct ether_vlan_header *)mp->b_rptr;
sap = ntohs(evhp->ether_type);
(void) mac_sap_verify(mh, sap, &mhip->mhi_bindsap);
mhip->mhi_hdrsize = sizeof (struct ether_vlan_header);
mhip->mhi_tci = ntohs(evhp->ether_tci);
mhip->mhi_istagged = B_TRUE;
freemsg(tmp);
if (VLAN_CFI(mhip->mhi_tci) != ETHER_CFI)
return (EINVAL);
} else {
mhip->mhi_istagged = B_FALSE;
mhip->mhi_tci = 0;
}
return (0);
}
mblk_t *
mac_header_cook(mac_handle_t mh, mblk_t *mp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
if (mip->mi_type->mt_ops.mtops_ops & MTOPS_HEADER_COOK) {
if (DB_REF(mp) > 1) {
mblk_t *newmp = copymsg(mp);
if (newmp == NULL)
return (NULL);
freemsg(mp);
mp = newmp;
}
return (mip->mi_type->mt_ops.mtops_header_cook(mp,
mip->mi_pdata));
}
return (mp);
}
mblk_t *
mac_header_uncook(mac_handle_t mh, mblk_t *mp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
if (mip->mi_type->mt_ops.mtops_ops & MTOPS_HEADER_UNCOOK) {
if (DB_REF(mp) > 1) {
mblk_t *newmp = copymsg(mp);
if (newmp == NULL)
return (NULL);
freemsg(mp);
mp = newmp;
}
return (mip->mi_type->mt_ops.mtops_header_uncook(mp,
mip->mi_pdata));
}
return (mp);
}
uint_t
mac_addr_len(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_type->mt_addr_length);
}
/* True if a MAC is a VNIC */
boolean_t
mac_is_vnic(mac_handle_t mh)
{
return (((mac_impl_t *)mh)->mi_state_flags & MIS_IS_VNIC);
}
mac_handle_t
mac_get_lower_mac_handle(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
ASSERT(mac_is_vnic(mh));
return (((vnic_t *)mip->mi_driver)->vn_lower_mh);
}
boolean_t
mac_is_vnic_primary(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
ASSERT(mac_is_vnic(mh));
return (((vnic_t *)mip->mi_driver)->vn_addr_type ==
VNIC_MAC_ADDR_TYPE_PRIMARY);
}
void
mac_update_resources(mac_resource_props_t *nmrp, mac_resource_props_t *cmrp,
boolean_t is_user_flow)
{
if (nmrp != NULL && cmrp != NULL) {
if (nmrp->mrp_mask & MRP_PRIORITY) {
if (nmrp->mrp_priority == MPL_RESET) {
cmrp->mrp_mask &= ~MRP_PRIORITY;
if (is_user_flow) {
cmrp->mrp_priority =
MPL_SUBFLOW_DEFAULT;
} else {
cmrp->mrp_priority = MPL_LINK_DEFAULT;
}
} else {
cmrp->mrp_mask |= MRP_PRIORITY;
cmrp->mrp_priority = nmrp->mrp_priority;
}
}
if (nmrp->mrp_mask & MRP_MAXBW) {
if (nmrp->mrp_maxbw == MRP_MAXBW_RESETVAL) {
cmrp->mrp_mask &= ~MRP_MAXBW;
cmrp->mrp_maxbw = 0;
} else {
cmrp->mrp_mask |= MRP_MAXBW;
cmrp->mrp_maxbw = nmrp->mrp_maxbw;
}
}
if (nmrp->mrp_mask & MRP_CPUS)
MAC_COPY_CPUS(nmrp, cmrp);
if (nmrp->mrp_mask & MRP_POOL) {
if (strlen(nmrp->mrp_pool) == 0) {
cmrp->mrp_mask &= ~MRP_POOL;
bzero(cmrp->mrp_pool, sizeof (cmrp->mrp_pool));
} else {
cmrp->mrp_mask |= MRP_POOL;
(void) strncpy(cmrp->mrp_pool, nmrp->mrp_pool,
sizeof (cmrp->mrp_pool));
}
}
if (nmrp->mrp_mask & MRP_PROTECT)
mac_protect_update(nmrp, cmrp);
/*
* Update the rings specified.
*/
if (nmrp->mrp_mask & MRP_RX_RINGS) {
if (nmrp->mrp_mask & MRP_RINGS_RESET) {
cmrp->mrp_mask &= ~MRP_RX_RINGS;
if (cmrp->mrp_mask & MRP_RXRINGS_UNSPEC)
cmrp->mrp_mask &= ~MRP_RXRINGS_UNSPEC;
cmrp->mrp_nrxrings = 0;
} else {
cmrp->mrp_mask |= MRP_RX_RINGS;
cmrp->mrp_nrxrings = nmrp->mrp_nrxrings;
}
}
if (nmrp->mrp_mask & MRP_TX_RINGS) {
if (nmrp->mrp_mask & MRP_RINGS_RESET) {
cmrp->mrp_mask &= ~MRP_TX_RINGS;
if (cmrp->mrp_mask & MRP_TXRINGS_UNSPEC)
cmrp->mrp_mask &= ~MRP_TXRINGS_UNSPEC;
cmrp->mrp_ntxrings = 0;
} else {
cmrp->mrp_mask |= MRP_TX_RINGS;
cmrp->mrp_ntxrings = nmrp->mrp_ntxrings;
}
}
if (nmrp->mrp_mask & MRP_RXRINGS_UNSPEC)
cmrp->mrp_mask |= MRP_RXRINGS_UNSPEC;
else if (cmrp->mrp_mask & MRP_RXRINGS_UNSPEC)
cmrp->mrp_mask &= ~MRP_RXRINGS_UNSPEC;
if (nmrp->mrp_mask & MRP_TXRINGS_UNSPEC)
cmrp->mrp_mask |= MRP_TXRINGS_UNSPEC;
else if (cmrp->mrp_mask & MRP_TXRINGS_UNSPEC)
cmrp->mrp_mask &= ~MRP_TXRINGS_UNSPEC;
}
}
/*
* i_mac_set_resources:
*
* This routine associates properties with the primary MAC client of
* the specified MAC instance.
* - Cache the properties in mac_impl_t
* - Apply the properties to the primary MAC client if exists
*/
int
i_mac_set_resources(mac_handle_t mh, mac_resource_props_t *mrp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *mcip;
int err = 0;
uint32_t resmask, newresmask;
mac_resource_props_t *tmrp, *umrp;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
err = mac_validate_props(mip, mrp);
if (err != 0)
return (err);
umrp = kmem_zalloc(sizeof (*umrp), KM_SLEEP);
bcopy(&mip->mi_resource_props, umrp, sizeof (*umrp));
resmask = umrp->mrp_mask;
mac_update_resources(mrp, umrp, B_FALSE);
newresmask = umrp->mrp_mask;
if (resmask == 0 && newresmask != 0) {
/*
* Bandwidth, priority, cpu or pool link properties configured,
* must disable fastpath.
*/
if ((err = mac_fastpath_disable((mac_handle_t)mip)) != 0) {
kmem_free(umrp, sizeof (*umrp));
return (err);
}
}
/*
* Since bind_cpu may be modified by mac_client_set_resources()
* we use a copy of bind_cpu and finally cache bind_cpu in mip.
* This allows us to cache only user edits in mip.
*/
tmrp = kmem_zalloc(sizeof (*tmrp), KM_SLEEP);
bcopy(mrp, tmrp, sizeof (*tmrp));
mcip = mac_primary_client_handle(mip);
if (mcip != NULL && (mcip->mci_state_flags & MCIS_IS_AGGR_PORT) == 0) {
err = mac_client_set_resources((mac_client_handle_t)mcip, tmrp);
} else if ((mrp->mrp_mask & MRP_RX_RINGS ||
mrp->mrp_mask & MRP_TX_RINGS)) {
mac_client_impl_t *vmcip;
/*
* If the primary is not up, we need to check if there
* are any VLANs on this primary. If there are then
* we need to set this property on the VLANs since
* VLANs follow the primary they are based on. Just
* look for the first VLAN and change its properties,
* all the other VLANs should be in the same group.
*/
for (vmcip = mip->mi_clients_list; vmcip != NULL;
vmcip = vmcip->mci_client_next) {
if ((vmcip->mci_flent->fe_type & FLOW_PRIMARY_MAC) &&
mac_client_vid((mac_client_handle_t)vmcip) !=
VLAN_ID_NONE) {
break;
}
}
if (vmcip != NULL) {
mac_resource_props_t *omrp;
mac_resource_props_t *vmrp;
omrp = kmem_zalloc(sizeof (*omrp), KM_SLEEP);
bcopy(MCIP_RESOURCE_PROPS(vmcip), omrp, sizeof (*omrp));
/*
* We dont' call mac_update_resources since we
* want to take only the ring properties and
* not all the properties that may have changed.
*/
vmrp = MCIP_RESOURCE_PROPS(vmcip);
if (mrp->mrp_mask & MRP_RX_RINGS) {
if (mrp->mrp_mask & MRP_RINGS_RESET) {
vmrp->mrp_mask &= ~MRP_RX_RINGS;
if (vmrp->mrp_mask &
MRP_RXRINGS_UNSPEC) {
vmrp->mrp_mask &=
~MRP_RXRINGS_UNSPEC;
}
vmrp->mrp_nrxrings = 0;
} else {
vmrp->mrp_mask |= MRP_RX_RINGS;
vmrp->mrp_nrxrings = mrp->mrp_nrxrings;
}
}
if (mrp->mrp_mask & MRP_TX_RINGS) {
if (mrp->mrp_mask & MRP_RINGS_RESET) {
vmrp->mrp_mask &= ~MRP_TX_RINGS;
if (vmrp->mrp_mask &
MRP_TXRINGS_UNSPEC) {
vmrp->mrp_mask &=
~MRP_TXRINGS_UNSPEC;
}
vmrp->mrp_ntxrings = 0;
} else {
vmrp->mrp_mask |= MRP_TX_RINGS;
vmrp->mrp_ntxrings = mrp->mrp_ntxrings;
}
}
if (mrp->mrp_mask & MRP_RXRINGS_UNSPEC)
vmrp->mrp_mask |= MRP_RXRINGS_UNSPEC;
if (mrp->mrp_mask & MRP_TXRINGS_UNSPEC)
vmrp->mrp_mask |= MRP_TXRINGS_UNSPEC;
if ((err = mac_client_set_rings_prop(vmcip, mrp,
omrp)) != 0) {
bcopy(omrp, MCIP_RESOURCE_PROPS(vmcip),
sizeof (*omrp));
} else {
mac_set_prim_vlan_rings(mip, vmrp);
}
kmem_free(omrp, sizeof (*omrp));
}
}
/* Only update the values if mac_client_set_resources succeeded */
if (err == 0) {
bcopy(umrp, &mip->mi_resource_props, sizeof (*umrp));
/*
* If bandwidth, priority or cpu link properties cleared,
* renable fastpath.
*/
if (resmask != 0 && newresmask == 0)
mac_fastpath_enable((mac_handle_t)mip);
} else if (resmask == 0 && newresmask != 0) {
mac_fastpath_enable((mac_handle_t)mip);
}
kmem_free(tmrp, sizeof (*tmrp));
kmem_free(umrp, sizeof (*umrp));
return (err);
}
int
mac_set_resources(mac_handle_t mh, mac_resource_props_t *mrp)
{
int err;
i_mac_perim_enter((mac_impl_t *)mh);
err = i_mac_set_resources(mh, mrp);
i_mac_perim_exit((mac_impl_t *)mh);
return (err);
}
/*
* Get the properties cached for the specified MAC instance.
*/
void
mac_get_resources(mac_handle_t mh, mac_resource_props_t *mrp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *mcip;
mcip = mac_primary_client_handle(mip);
if (mcip != NULL) {
mac_client_get_resources((mac_client_handle_t)mcip, mrp);
return;
}
bcopy(&mip->mi_resource_props, mrp, sizeof (mac_resource_props_t));
}
/*
* Get the effective properties from the primary client of the
* specified MAC instance.
*/
void
mac_get_effective_resources(mac_handle_t mh, mac_resource_props_t *mrp)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *mcip;
mcip = mac_primary_client_handle(mip);
if (mcip != NULL) {
mac_client_get_effective_resources((mac_client_handle_t)mcip,
mrp);
return;
}
bzero(mrp, sizeof (mac_resource_props_t));
}
int
mac_set_pvid(mac_handle_t mh, uint16_t pvid)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *mcip;
mac_unicast_impl_t *muip;
i_mac_perim_enter(mip);
if (pvid != 0) {
for (mcip = mip->mi_clients_list; mcip != NULL;
mcip = mcip->mci_client_next) {
for (muip = mcip->mci_unicast_list; muip != NULL;
muip = muip->mui_next) {
if (muip->mui_vid == pvid) {
i_mac_perim_exit(mip);
return (EBUSY);
}
}
}
}
mip->mi_pvid = pvid;
i_mac_perim_exit(mip);
return (0);
}
uint16_t
mac_get_pvid(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_pvid);
}
uint32_t
mac_get_llimit(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_llimit);
}
uint32_t
mac_get_ldecay(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_ldecay);
}
/*
* Rename a mac client, its flow, and the kstat.
*/
int
mac_rename_primary(mac_handle_t mh, const char *new_name)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_client_impl_t *cur_clnt = NULL;
flow_entry_t *fep;
i_mac_perim_enter(mip);
/*
* VNICs: we need to change the sys flow name and
* the associated flow kstat.
*/
if (mip->mi_state_flags & MIS_IS_VNIC) {
mac_client_impl_t *mcip = mac_vnic_lower(mip);
ASSERT(new_name != NULL);
mac_rename_flow_names(mcip, new_name);
mac_stat_rename(mcip);
goto done;
}
/*
* This mac may itself be an aggr link, or it may have some client
* which is an aggr port. For both cases, we need to change the
* aggr port's mac client name, its flow name and the associated flow
* kstat.
*/
if (mip->mi_state_flags & MIS_IS_AGGR) {
mac_capab_aggr_t aggr_cap;
mac_rename_fn_t rename_fn;
boolean_t ret;
ASSERT(new_name != NULL);
ret = i_mac_capab_get((mac_handle_t)mip, MAC_CAPAB_AGGR,
(void *)(&aggr_cap));
ASSERT(ret == B_TRUE);
rename_fn = aggr_cap.mca_rename_fn;
rename_fn(new_name, mip->mi_driver);
/*
* The aggr's client name and kstat flow name will be
* updated below, i.e. via mac_rename_flow_names.
*/
}
for (cur_clnt = mip->mi_clients_list; cur_clnt != NULL;
cur_clnt = cur_clnt->mci_client_next) {
if (cur_clnt->mci_state_flags & MCIS_IS_AGGR_PORT) {
if (new_name != NULL) {
char *str_st = cur_clnt->mci_name;
char *str_del = strchr(str_st, '-');
ASSERT(str_del != NULL);
bzero(str_del + 1, MAXNAMELEN -
(str_del - str_st + 1));
bcopy(new_name, str_del + 1,
strlen(new_name));
}
fep = cur_clnt->mci_flent;
mac_rename_flow(fep, cur_clnt->mci_name);
break;
} else if (new_name != NULL &&
cur_clnt->mci_state_flags & MCIS_USE_DATALINK_NAME) {
mac_rename_flow_names(cur_clnt, new_name);
break;
}
}
/* Recreate kstats associated with aggr pseudo rings */
if (mip->mi_state_flags & MIS_IS_AGGR)
mac_pseudo_ring_stat_rename(mip);
done:
i_mac_perim_exit(mip);
return (0);
}
/*
* Rename the MAC client's flow names
*/
static void
mac_rename_flow_names(mac_client_impl_t *mcip, const char *new_name)
{
flow_entry_t *flent;
uint16_t vid;
char flowname[MAXFLOWNAMELEN];
mac_impl_t *mip = mcip->mci_mip;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mip));
/*
* Use mi_rw_lock to ensure that threads not in the mac perimeter
* see a self-consistent value for mci_name
*/
rw_enter(&mip->mi_rw_lock, RW_WRITER);
(void) strlcpy(mcip->mci_name, new_name, sizeof (mcip->mci_name));
rw_exit(&mip->mi_rw_lock);
mac_rename_flow(mcip->mci_flent, new_name);
if (mcip->mci_nflents == 1)
return;
/*
* We have to rename all the others too, no stats to destroy for
* these.
*/
for (flent = mcip->mci_flent_list; flent != NULL;
flent = flent->fe_client_next) {
if (flent != mcip->mci_flent) {
vid = i_mac_flow_vid(flent);
(void) sprintf(flowname, "%s%u", new_name, vid);
mac_flow_set_name(flent, flowname);
}
}
}
/*
* Add a flow to the MAC client's flow list - i.e list of MAC/VID tuples
* defined for the specified MAC client.
*/
static void
mac_client_add_to_flow_list(mac_client_impl_t *mcip, flow_entry_t *flent)
{
ASSERT(MAC_PERIM_HELD((mac_handle_t)mcip->mci_mip));
/*
* The promisc Rx data path walks the mci_flent_list. Protect by
* using mi_rw_lock
*/
rw_enter(&mcip->mci_rw_lock, RW_WRITER);
mcip->mci_vidcache = MCIP_VIDCACHE_INVALID;
/* Add it to the head */
flent->fe_client_next = mcip->mci_flent_list;
mcip->mci_flent_list = flent;
mcip->mci_nflents++;
/*
* Keep track of the number of non-zero VIDs addresses per MAC
* client to avoid figuring it out in the data-path.
*/
if (i_mac_flow_vid(flent) != VLAN_ID_NONE)
mcip->mci_nvids++;
rw_exit(&mcip->mci_rw_lock);
}
/*
* Remove a flow entry from the MAC client's list.
*/
static void
mac_client_remove_flow_from_list(mac_client_impl_t *mcip, flow_entry_t *flent)
{
flow_entry_t *fe = mcip->mci_flent_list;
flow_entry_t *prev_fe = NULL;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mcip->mci_mip));
/*
* The promisc Rx data path walks the mci_flent_list. Protect by
* using mci_rw_lock
*/
rw_enter(&mcip->mci_rw_lock, RW_WRITER);
mcip->mci_vidcache = MCIP_VIDCACHE_INVALID;
while ((fe != NULL) && (fe != flent)) {
prev_fe = fe;
fe = fe->fe_client_next;
}
ASSERT(fe != NULL);
if (prev_fe == NULL) {
/* Deleting the first node */
mcip->mci_flent_list = fe->fe_client_next;
} else {
prev_fe->fe_client_next = fe->fe_client_next;
}
mcip->mci_nflents--;
if (i_mac_flow_vid(flent) != VLAN_ID_NONE)
mcip->mci_nvids--;
rw_exit(&mcip->mci_rw_lock);
}
/*
* Check if the given VID belongs to this MAC client.
*/
boolean_t
mac_client_check_flow_vid(mac_client_impl_t *mcip, uint16_t vid)
{
flow_entry_t *flent;
uint16_t mci_vid;
uint32_t cache = mcip->mci_vidcache;
/*
* In hopes of not having to touch the mci_rw_lock, check to see if
* this vid matches our cached result.
*/
if (MCIP_VIDCACHE_ISVALID(cache) && MCIP_VIDCACHE_VID(cache) == vid)
return (MCIP_VIDCACHE_BOOL(cache) ? B_TRUE : B_FALSE);
/* The mci_flent_list is protected by mci_rw_lock */
rw_enter(&mcip->mci_rw_lock, RW_WRITER);
for (flent = mcip->mci_flent_list; flent != NULL;
flent = flent->fe_client_next) {
mci_vid = i_mac_flow_vid(flent);
if (vid == mci_vid) {
mcip->mci_vidcache = MCIP_VIDCACHE_CACHE(vid, B_TRUE);
rw_exit(&mcip->mci_rw_lock);
return (B_TRUE);
}
}
mcip->mci_vidcache = MCIP_VIDCACHE_CACHE(vid, B_FALSE);
rw_exit(&mcip->mci_rw_lock);
return (B_FALSE);
}
/*
* Get the flow entry for the specified <MAC addr, VID> tuple.
*/
static flow_entry_t *
mac_client_get_flow(mac_client_impl_t *mcip, mac_unicast_impl_t *muip)
{
mac_address_t *map = mcip->mci_unicast;
flow_entry_t *flent;
uint16_t vid;
flow_desc_t flow_desc;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mcip->mci_mip));
mac_flow_get_desc(mcip->mci_flent, &flow_desc);
if (bcmp(flow_desc.fd_dst_mac, map->ma_addr, map->ma_len) != 0)
return (NULL);
for (flent = mcip->mci_flent_list; flent != NULL;
flent = flent->fe_client_next) {
vid = i_mac_flow_vid(flent);
if (vid == muip->mui_vid) {
return (flent);
}
}
return (NULL);
}
/*
* Since mci_flent has the SRSs, when we want to remove it, we replace
* the flow_desc_t in mci_flent with that of an existing flent and then
* remove that flent instead of mci_flent.
*/
static flow_entry_t *
mac_client_swap_mciflent(mac_client_impl_t *mcip)
{
flow_entry_t *flent = mcip->mci_flent;
flow_tab_t *ft = flent->fe_flow_tab;
flow_entry_t *flent1;
flow_desc_t fl_desc;
char fl_name[MAXFLOWNAMELEN];
int err;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mcip->mci_mip));
ASSERT(mcip->mci_nflents > 1);
/* get the next flent following the primary flent */
flent1 = mcip->mci_flent_list->fe_client_next;
ASSERT(flent1 != NULL && flent1->fe_flow_tab == ft);
/*
* Remove the flent from the flow table before updating the
* flow descriptor as the hash depends on the flow descriptor.
* This also helps incoming packet classification avoid having
* to grab fe_lock. Access to fe_flow_desc of a flent not in the
* flow table is done under the fe_lock so that log or stat functions
* see a self-consistent fe_flow_desc. The name and desc are specific
* to a flow, the rest are shared by all the clients, including
* resource control etc.
*/
mac_flow_remove(ft, flent, B_TRUE);
mac_flow_remove(ft, flent1, B_TRUE);
bcopy(&flent->fe_flow_desc, &fl_desc, sizeof (flow_desc_t));
bcopy(flent->fe_flow_name, fl_name, MAXFLOWNAMELEN);
/* update the primary flow entry */
mutex_enter(&flent->fe_lock);
bcopy(&flent1->fe_flow_desc, &flent->fe_flow_desc,
sizeof (flow_desc_t));
bcopy(&flent1->fe_flow_name, &flent->fe_flow_name, MAXFLOWNAMELEN);
mutex_exit(&flent->fe_lock);
/* update the flow entry that is to be freed */
mutex_enter(&flent1->fe_lock);
bcopy(&fl_desc, &flent1->fe_flow_desc, sizeof (flow_desc_t));
bcopy(fl_name, &flent1->fe_flow_name, MAXFLOWNAMELEN);
mutex_exit(&flent1->fe_lock);
/* now reinsert the flow entries in the table */
err = mac_flow_add(ft, flent);
ASSERT(err == 0);
err = mac_flow_add(ft, flent1);
ASSERT(err == 0);
return (flent1);
}
/*
* Return whether there is only one flow entry associated with this
* MAC client.
*/
static boolean_t
mac_client_single_rcvr(mac_client_impl_t *mcip)
{
return (mcip->mci_nflents == 1);
}
int
mac_validate_props(mac_impl_t *mip, mac_resource_props_t *mrp)
{
boolean_t reset;
uint32_t rings_needed;
uint32_t rings_avail;
mac_group_type_t gtype;
mac_resource_props_t *mip_mrp;
if (mrp == NULL)
return (0);
if (mrp->mrp_mask & MRP_PRIORITY) {
mac_priority_level_t pri = mrp->mrp_priority;
if (pri < MPL_LOW || pri > MPL_RESET)
return (EINVAL);
}
if (mrp->mrp_mask & MRP_MAXBW) {
uint64_t maxbw = mrp->mrp_maxbw;
if (maxbw < MRP_MAXBW_MINVAL && maxbw != 0)
return (EINVAL);
}
if (mrp->mrp_mask & MRP_CPUS) {
int i, j;
mac_cpu_mode_t fanout;
if (mrp->mrp_ncpus > ncpus)
return (EINVAL);
for (i = 0; i < mrp->mrp_ncpus; i++) {
for (j = 0; j < mrp->mrp_ncpus; j++) {
if (i != j &&
mrp->mrp_cpu[i] == mrp->mrp_cpu[j]) {
return (EINVAL);
}
}
}
for (i = 0; i < mrp->mrp_ncpus; i++) {
cpu_t *cp;
int rv;
mutex_enter(&cpu_lock);
cp = cpu_get(mrp->mrp_cpu[i]);
if (cp != NULL)
rv = cpu_is_online(cp);
else
rv = 0;
mutex_exit(&cpu_lock);
if (rv == 0)
return (EINVAL);
}
fanout = mrp->mrp_fanout_mode;
if (fanout < 0 || fanout > MCM_CPUS)
return (EINVAL);
}
if (mrp->mrp_mask & MRP_PROTECT) {
int err = mac_protect_validate(mrp);
if (err != 0)
return (err);
}
if (!(mrp->mrp_mask & MRP_RX_RINGS) &&
!(mrp->mrp_mask & MRP_TX_RINGS)) {
return (0);
}
/*
* mip will be null when we come from mac_flow_create or
* mac_link_flow_modify. In the latter case it is a user flow,
* for which we don't support rings. In the former we would
* have validated the props beforehand (i_mac_unicast_add ->
* mac_client_set_resources -> validate for the primary and
* vnic_dev_create -> mac_client_set_resources -> validate for
* a vnic.
*/
if (mip == NULL)
return (0);
/*
* We don't support setting rings property for a VNIC that is using a
* primary address (VLAN)
*/
if ((mip->mi_state_flags & MIS_IS_VNIC) &&
mac_is_vnic_primary((mac_handle_t)mip)) {
return (ENOTSUP);
}
mip_mrp = &mip->mi_resource_props;
/*
* The rings property should be validated against the NICs
* resources
*/
if (mip->mi_state_flags & MIS_IS_VNIC)
mip = (mac_impl_t *)mac_get_lower_mac_handle((mac_handle_t)mip);
reset = mrp->mrp_mask & MRP_RINGS_RESET;
/*
* If groups are not supported, return error.
*/
if (((mrp->mrp_mask & MRP_RX_RINGS) && mip->mi_rx_groups == NULL) ||
((mrp->mrp_mask & MRP_TX_RINGS) && mip->mi_tx_groups == NULL)) {
return (EINVAL);
}
/*
* If we are just resetting, there is no validation needed.
*/
if (reset)
return (0);
if (mrp->mrp_mask & MRP_RX_RINGS) {
rings_needed = mrp->mrp_nrxrings;
/*
* We just want to check if the number of additional
* rings requested is available.
*/
if (mip_mrp->mrp_mask & MRP_RX_RINGS) {
if (mrp->mrp_nrxrings > mip_mrp->mrp_nrxrings)
/* Just check for the additional rings */
rings_needed -= mip_mrp->mrp_nrxrings;
else
/* We are not asking for additional rings */
rings_needed = 0;
}
rings_avail = mip->mi_rxrings_avail;
gtype = mip->mi_rx_group_type;
} else {
rings_needed = mrp->mrp_ntxrings;
/* Similarly for the TX rings */
if (mip_mrp->mrp_mask & MRP_TX_RINGS) {
if (mrp->mrp_ntxrings > mip_mrp->mrp_ntxrings)
/* Just check for the additional rings */
rings_needed -= mip_mrp->mrp_ntxrings;
else
/* We are not asking for additional rings */
rings_needed = 0;
}
rings_avail = mip->mi_txrings_avail;
gtype = mip->mi_tx_group_type;
}
/* Error if the group is dynamic .. */
if (gtype == MAC_GROUP_TYPE_DYNAMIC) {
/*
* .. and rings specified are more than available.
*/
if (rings_needed > rings_avail)
return (EINVAL);
} else {
/*
* OR group is static and we have specified some rings.
*/
if (rings_needed > 0)
return (EINVAL);
}
return (0);
}
/*
* Send a MAC_NOTE_LINK notification to all the MAC clients whenever the
* underlying physical link is down. This is to allow MAC clients to
* communicate with other clients.
*/
void
mac_virtual_link_update(mac_impl_t *mip)
{
if (mip->mi_linkstate != LINK_STATE_UP)
i_mac_notify(mip, MAC_NOTE_LINK);
}
/*
* For clients that have a pass-thru MAC, e.g. VNIC, we set the VNIC's
* mac handle in the client.
*/
void
mac_set_upper_mac(mac_client_handle_t mch, mac_handle_t mh,
mac_resource_props_t *mrp)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_impl_t *mip = (mac_impl_t *)mh;
mcip->mci_upper_mip = mip;
/* If there are any properties, copy it over too */
if (mrp != NULL) {
bcopy(mrp, &mip->mi_resource_props,
sizeof (mac_resource_props_t));
}
}
/*
* Mark the mac as being used exclusively by the single mac client that is
* doing some control operation on this mac. No further opens of this mac
* will be allowed until this client calls mac_unmark_exclusive. The mac
* client calling this function must already be in the mac perimeter
*/
int
mac_mark_exclusive(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
ASSERT(MAC_PERIM_HELD(mh));
/*
* Look up its entry in the global hash table.
*/
rw_enter(&i_mac_impl_lock, RW_WRITER);
if (mip->mi_state_flags & MIS_DISABLED) {
rw_exit(&i_mac_impl_lock);
return (ENOENT);
}
/*
* A reference to mac is held even if the link is not plumbed.
* In i_dls_link_create() we open the MAC interface and hold the
* reference. There is an additional reference for the mac_open
* done in acquiring the mac perimeter
*/
if (mip->mi_ref != 2) {
rw_exit(&i_mac_impl_lock);
return (EBUSY);
}
ASSERT(!(mip->mi_state_flags & MIS_EXCLUSIVE_HELD));
mip->mi_state_flags |= MIS_EXCLUSIVE_HELD;
rw_exit(&i_mac_impl_lock);
return (0);
}
void
mac_unmark_exclusive(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
ASSERT(MAC_PERIM_HELD(mh));
rw_enter(&i_mac_impl_lock, RW_WRITER);
/* 1 for the creation and another for the perimeter */
ASSERT(mip->mi_ref == 2 && (mip->mi_state_flags & MIS_EXCLUSIVE_HELD));
mip->mi_state_flags &= ~MIS_EXCLUSIVE_HELD;
rw_exit(&i_mac_impl_lock);
}
/*
* Set the MTU for the specified MAC.
*/
int
mac_set_mtu(mac_handle_t mh, uint_t new_mtu, uint_t *old_mtu_arg)
{
mac_impl_t *mip = (mac_impl_t *)mh;
uint_t old_mtu;
int rv = 0;
i_mac_perim_enter(mip);
if (!(mip->mi_callbacks->mc_callbacks & (MC_SETPROP|MC_GETPROP))) {
rv = ENOTSUP;
goto bail;
}
old_mtu = mip->mi_sdu_max;
if (new_mtu == 0 || new_mtu < mip->mi_sdu_min) {
rv = EINVAL;
goto bail;
}
rw_enter(&mip->mi_rw_lock, RW_READER);
if (mip->mi_mtrp != NULL && new_mtu < mip->mi_mtrp->mtr_mtu) {
rv = EBUSY;
rw_exit(&mip->mi_rw_lock);
goto bail;
}
rw_exit(&mip->mi_rw_lock);
if (old_mtu != new_mtu) {
rv = mip->mi_callbacks->mc_setprop(mip->mi_driver,
"mtu", MAC_PROP_MTU, sizeof (uint_t), &new_mtu);
if (rv != 0)
goto bail;
rv = mac_maxsdu_update(mh, new_mtu);
ASSERT(rv == 0);
}
bail:
i_mac_perim_exit(mip);
if (rv == 0 && old_mtu_arg != NULL)
*old_mtu_arg = old_mtu;
return (rv);
}
/*
* Return the RX h/w information for the group indexed by grp_num.
*/
void
mac_get_hwrxgrp_info(mac_handle_t mh, int grp_index, uint_t *grp_num,
uint_t *n_rings, uint_t *rings, uint_t *type, uint_t *n_clnts,
char *clnts_name)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_grp_client_t *mcip;
uint_t i = 0, index = 0;
mac_ring_t *ring;
/* Revisit when we implement fully dynamic group allocation */
ASSERT(grp_index >= 0 && grp_index < mip->mi_rx_group_count);
rw_enter(&mip->mi_rw_lock, RW_READER);
*grp_num = mip->mi_rx_groups[grp_index].mrg_index;
*type = mip->mi_rx_groups[grp_index].mrg_type;
*n_rings = mip->mi_rx_groups[grp_index].mrg_cur_count;
ring = mip->mi_rx_groups[grp_index].mrg_rings;
for (index = 0; index < mip->mi_rx_groups[grp_index].mrg_cur_count;
index++) {
rings[index] = ring->mr_index;
ring = ring->mr_next;
}
/* Assuming the 1st is the default group */
index = 0;
if (grp_index == 0) {
(void) strlcpy(clnts_name, "<default,mcast>,",
MAXCLIENTNAMELEN);
index += strlen("<default,mcast>,");
}
for (mcip = mip->mi_rx_groups[grp_index].mrg_clients; mcip != NULL;
mcip = mcip->mgc_next) {
int name_len = strlen(mcip->mgc_client->mci_name);
/*
* MAXCLIENTNAMELEN is the buffer size reserved for client
* names.
* XXXX Formating the client name string needs to be moved
* to user land when fixing the size of dhi_clnts in
* dld_hwgrpinfo_t. We should use n_clients * client_name for
* dhi_clntsin instead of MAXCLIENTNAMELEN
*/
if (index + name_len >= MAXCLIENTNAMELEN) {
index = MAXCLIENTNAMELEN;
break;
}
bcopy(mcip->mgc_client->mci_name, &(clnts_name[index]),
name_len);
index += name_len;
clnts_name[index++] = ',';
i++;
}
/* Get rid of the last , */
if (index > 0)
clnts_name[index - 1] = '\0';
*n_clnts = i;
rw_exit(&mip->mi_rw_lock);
}
/*
* Return the TX h/w information for the group indexed by grp_num.
*/
void
mac_get_hwtxgrp_info(mac_handle_t mh, int grp_index, uint_t *grp_num,
uint_t *n_rings, uint_t *rings, uint_t *type, uint_t *n_clnts,
char *clnts_name)
{
mac_impl_t *mip = (mac_impl_t *)mh;
mac_grp_client_t *mcip;
uint_t i = 0, index = 0;
mac_ring_t *ring;
/* Revisit when we implement fully dynamic group allocation */
ASSERT(grp_index >= 0 && grp_index <= mip->mi_tx_group_count);
rw_enter(&mip->mi_rw_lock, RW_READER);
*grp_num = mip->mi_tx_groups[grp_index].mrg_index > 0 ?
mip->mi_tx_groups[grp_index].mrg_index : grp_index;
*type = mip->mi_tx_groups[grp_index].mrg_type;
*n_rings = mip->mi_tx_groups[grp_index].mrg_cur_count;
ring = mip->mi_tx_groups[grp_index].mrg_rings;
for (index = 0; index < mip->mi_tx_groups[grp_index].mrg_cur_count;
index++) {
rings[index] = ring->mr_index;
ring = ring->mr_next;
}
index = 0;
/* Default group has an index of -1 */
if (mip->mi_tx_groups[grp_index].mrg_index < 0) {
(void) strlcpy(clnts_name, "<default>,",
MAXCLIENTNAMELEN);
index += strlen("<default>,");
}
for (mcip = mip->mi_tx_groups[grp_index].mrg_clients; mcip != NULL;
mcip = mcip->mgc_next) {
int name_len = strlen(mcip->mgc_client->mci_name);
/*
* MAXCLIENTNAMELEN is the buffer size reserved for client
* names.
* XXXX Formating the client name string needs to be moved
* to user land when fixing the size of dhi_clnts in
* dld_hwgrpinfo_t. We should use n_clients * client_name for
* dhi_clntsin instead of MAXCLIENTNAMELEN
*/
if (index + name_len >= MAXCLIENTNAMELEN) {
index = MAXCLIENTNAMELEN;
break;
}
bcopy(mcip->mgc_client->mci_name, &(clnts_name[index]),
name_len);
index += name_len;
clnts_name[index++] = ',';
i++;
}
/* Get rid of the last , */
if (index > 0)
clnts_name[index - 1] = '\0';
*n_clnts = i;
rw_exit(&mip->mi_rw_lock);
}
/*
* Return the group count for RX or TX.
*/
uint_t
mac_hwgrp_num(mac_handle_t mh, int type)
{
mac_impl_t *mip = (mac_impl_t *)mh;
/*
* Return the Rx and Tx group count; for the Tx we need to
* include the default too.
*/
return (type == MAC_RING_TYPE_RX ? mip->mi_rx_group_count :
mip->mi_tx_groups != NULL ? mip->mi_tx_group_count + 1 : 0);
}
/*
* The total number of free TX rings for this MAC.
*/
uint_t
mac_txavail_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_txrings_avail);
}
/*
* The total number of free RX rings for this MAC.
*/
uint_t
mac_rxavail_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_rxrings_avail);
}
/*
* The total number of reserved RX rings on this MAC.
*/
uint_t
mac_rxrsvd_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_rxrings_rsvd);
}
/*
* The total number of reserved TX rings on this MAC.
*/
uint_t
mac_txrsvd_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_txrings_rsvd);
}
/*
* Total number of free RX groups on this MAC.
*/
uint_t
mac_rxhwlnksavail_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_rxhwclnt_avail);
}
/*
* Total number of RX groups reserved on this MAC.
*/
uint_t
mac_rxhwlnksrsvd_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_rxhwclnt_used);
}
/*
* Total number of free TX groups on this MAC.
*/
uint_t
mac_txhwlnksavail_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_txhwclnt_avail);
}
/*
* Total number of TX groups reserved on this MAC.
*/
uint_t
mac_txhwlnksrsvd_get(mac_handle_t mh)
{
mac_impl_t *mip = (mac_impl_t *)mh;
return (mip->mi_txhwclnt_used);
}
/*
* Initialize the rings property for a mac client. A non-0 value for
* rxring or txring specifies the number of rings required, a value
* of MAC_RXRINGS_NONE/MAC_TXRINGS_NONE specifies that it doesn't need
* any RX/TX rings and a value of MAC_RXRINGS_DONTCARE/MAC_TXRINGS_DONTCARE
* means the system can decide whether it can give any rings or not.
*/
void
mac_client_set_rings(mac_client_handle_t mch, int rxrings, int txrings)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
mac_resource_props_t *mrp = MCIP_RESOURCE_PROPS(mcip);
if (rxrings != MAC_RXRINGS_DONTCARE) {
mrp->mrp_mask |= MRP_RX_RINGS;
mrp->mrp_nrxrings = rxrings;
}
if (txrings != MAC_TXRINGS_DONTCARE) {
mrp->mrp_mask |= MRP_TX_RINGS;
mrp->mrp_ntxrings = txrings;
}
}
boolean_t
mac_get_promisc_filtered(mac_client_handle_t mch)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
return (mcip->mci_protect_flags & MPT_FLAG_PROMISC_FILTERED);
}
void
mac_set_promisc_filtered(mac_client_handle_t mch, boolean_t enable)
{
mac_client_impl_t *mcip = (mac_client_impl_t *)mch;
ASSERT(MAC_PERIM_HELD((mac_handle_t)mcip->mci_mip));
if (enable)
mcip->mci_protect_flags |= MPT_FLAG_PROMISC_FILTERED;
else
mcip->mci_protect_flags &= ~MPT_FLAG_PROMISC_FILTERED;
}