242 lines
		
	
	
		
			5.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			242 lines
		
	
	
		
			5.5 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /* Copyright (c) 2019 Chelsio Communications, Inc. All rights reserved. */
 | |
| 
 | |
| #include "cxgb4.h"
 | |
| 
 | |
| static int cxgb4_mps_ref_dec_by_mac(struct adapter *adap,
 | |
| 				    const u8 *addr, const u8 *mask)
 | |
| {
 | |
| 	u8 bitmask[] = { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
 | |
| 	struct mps_entries_ref *mps_entry, *tmp;
 | |
| 	int ret = -EINVAL;
 | |
| 
 | |
| 	spin_lock_bh(&adap->mps_ref_lock);
 | |
| 	list_for_each_entry_safe(mps_entry, tmp, &adap->mps_ref, list) {
 | |
| 		if (ether_addr_equal(mps_entry->addr, addr) &&
 | |
| 		    ether_addr_equal(mps_entry->mask, mask ? mask : bitmask)) {
 | |
| 			if (!refcount_dec_and_test(&mps_entry->refcnt)) {
 | |
| 				spin_unlock_bh(&adap->mps_ref_lock);
 | |
| 				return -EBUSY;
 | |
| 			}
 | |
| 			list_del(&mps_entry->list);
 | |
| 			kfree(mps_entry);
 | |
| 			ret = 0;
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock_bh(&adap->mps_ref_lock);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int cxgb4_mps_ref_dec(struct adapter *adap, u16 idx)
 | |
| {
 | |
| 	struct mps_entries_ref *mps_entry, *tmp;
 | |
| 	int ret = -EINVAL;
 | |
| 
 | |
| 	spin_lock(&adap->mps_ref_lock);
 | |
| 	list_for_each_entry_safe(mps_entry, tmp, &adap->mps_ref, list) {
 | |
| 		if (mps_entry->idx == idx) {
 | |
| 			if (!refcount_dec_and_test(&mps_entry->refcnt)) {
 | |
| 				spin_unlock(&adap->mps_ref_lock);
 | |
| 				return -EBUSY;
 | |
| 			}
 | |
| 			list_del(&mps_entry->list);
 | |
| 			kfree(mps_entry);
 | |
| 			ret = 0;
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	spin_unlock(&adap->mps_ref_lock);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int cxgb4_mps_ref_inc(struct adapter *adap, const u8 *mac_addr,
 | |
| 			     u16 idx, const u8 *mask)
 | |
| {
 | |
| 	u8 bitmask[] = { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
 | |
| 	struct mps_entries_ref *mps_entry;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	spin_lock_bh(&adap->mps_ref_lock);
 | |
| 	list_for_each_entry(mps_entry, &adap->mps_ref, list) {
 | |
| 		if (mps_entry->idx == idx) {
 | |
| 			refcount_inc(&mps_entry->refcnt);
 | |
| 			goto unlock;
 | |
| 		}
 | |
| 	}
 | |
| 	mps_entry = kzalloc(sizeof(*mps_entry), GFP_ATOMIC);
 | |
| 	if (!mps_entry) {
 | |
| 		ret = -ENOMEM;
 | |
| 		goto unlock;
 | |
| 	}
 | |
| 	ether_addr_copy(mps_entry->mask, mask ? mask : bitmask);
 | |
| 	ether_addr_copy(mps_entry->addr, mac_addr);
 | |
| 	mps_entry->idx = idx;
 | |
| 	refcount_set(&mps_entry->refcnt, 1);
 | |
| 	list_add_tail(&mps_entry->list, &adap->mps_ref);
 | |
| unlock:
 | |
| 	spin_unlock_bh(&adap->mps_ref_lock);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_free_mac_filt(struct adapter *adap, unsigned int viid,
 | |
| 			unsigned int naddr, const u8 **addr, bool sleep_ok)
 | |
| {
 | |
| 	int ret, i;
 | |
| 
 | |
| 	for (i = 0; i < naddr; i++) {
 | |
| 		if (!cxgb4_mps_ref_dec_by_mac(adap, addr[i], NULL)) {
 | |
| 			ret = t4_free_mac_filt(adap, adap->mbox, viid,
 | |
| 					       1, &addr[i], sleep_ok);
 | |
| 			if (ret < 0)
 | |
| 				return ret;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* return number of filters freed */
 | |
| 	return naddr;
 | |
| }
 | |
| 
 | |
| int cxgb4_alloc_mac_filt(struct adapter *adap, unsigned int viid,
 | |
| 			 bool free, unsigned int naddr, const u8 **addr,
 | |
| 			 u16 *idx, u64 *hash, bool sleep_ok)
 | |
| {
 | |
| 	int ret, i;
 | |
| 
 | |
| 	ret = t4_alloc_mac_filt(adap, adap->mbox, viid, free,
 | |
| 				naddr, addr, idx, hash, sleep_ok);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	for (i = 0; i < naddr; i++) {
 | |
| 		if (idx[i] != 0xffff) {
 | |
| 			if (cxgb4_mps_ref_inc(adap, addr[i], idx[i], NULL)) {
 | |
| 				ret = -ENOMEM;
 | |
| 				goto error;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	goto out;
 | |
| error:
 | |
| 	cxgb4_free_mac_filt(adap, viid, naddr, addr, sleep_ok);
 | |
| 
 | |
| out:
 | |
| 	/* Returns a negative error number or the number of filters allocated */
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_update_mac_filt(struct port_info *pi, unsigned int viid,
 | |
| 			  int *tcam_idx, const u8 *addr,
 | |
| 			  bool persistent, u8 *smt_idx)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = cxgb4_change_mac(pi, viid, tcam_idx,
 | |
| 			       addr, persistent, smt_idx);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	cxgb4_mps_ref_inc(pi->adapter, addr, *tcam_idx, NULL);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_free_raw_mac_filt(struct adapter *adap,
 | |
| 			    unsigned int viid,
 | |
| 			    const u8 *addr,
 | |
| 			    const u8 *mask,
 | |
| 			    unsigned int idx,
 | |
| 			    u8 lookup_type,
 | |
| 			    u8 port_id,
 | |
| 			    bool sleep_ok)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	if (!cxgb4_mps_ref_dec(adap, idx))
 | |
| 		ret = t4_free_raw_mac_filt(adap, viid, addr,
 | |
| 					   mask, idx, lookup_type,
 | |
| 					   port_id, sleep_ok);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_alloc_raw_mac_filt(struct adapter *adap,
 | |
| 			     unsigned int viid,
 | |
| 			     const u8 *addr,
 | |
| 			     const u8 *mask,
 | |
| 			     unsigned int idx,
 | |
| 			     u8 lookup_type,
 | |
| 			     u8 port_id,
 | |
| 			     bool sleep_ok)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = t4_alloc_raw_mac_filt(adap, viid, addr,
 | |
| 				    mask, idx, lookup_type,
 | |
| 				    port_id, sleep_ok);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (cxgb4_mps_ref_inc(adap, addr, ret, mask)) {
 | |
| 		ret = -ENOMEM;
 | |
| 		t4_free_raw_mac_filt(adap, viid, addr,
 | |
| 				     mask, idx, lookup_type,
 | |
| 				     port_id, sleep_ok);
 | |
| 	}
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_free_encap_mac_filt(struct adapter *adap, unsigned int viid,
 | |
| 			      int idx, bool sleep_ok)
 | |
| {
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	if (!cxgb4_mps_ref_dec(adap, idx))
 | |
| 		ret = t4_free_encap_mac_filt(adap, viid, idx, sleep_ok);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_alloc_encap_mac_filt(struct adapter *adap, unsigned int viid,
 | |
| 			       const u8 *addr, const u8 *mask,
 | |
| 			       unsigned int vni, unsigned int vni_mask,
 | |
| 			       u8 dip_hit, u8 lookup_type, bool sleep_ok)
 | |
| {
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = t4_alloc_encap_mac_filt(adap, viid, addr, mask, vni, vni_mask,
 | |
| 				      dip_hit, lookup_type, sleep_ok);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (cxgb4_mps_ref_inc(adap, addr, ret, mask)) {
 | |
| 		ret = -ENOMEM;
 | |
| 		t4_free_encap_mac_filt(adap, viid, ret, sleep_ok);
 | |
| 	}
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int cxgb4_init_mps_ref_entries(struct adapter *adap)
 | |
| {
 | |
| 	spin_lock_init(&adap->mps_ref_lock);
 | |
| 	INIT_LIST_HEAD(&adap->mps_ref);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void cxgb4_free_mps_ref_entries(struct adapter *adap)
 | |
| {
 | |
| 	struct mps_entries_ref *mps_entry, *tmp;
 | |
| 
 | |
| 	if (list_empty(&adap->mps_ref))
 | |
| 		return;
 | |
| 
 | |
| 	spin_lock(&adap->mps_ref_lock);
 | |
| 	list_for_each_entry_safe(mps_entry, tmp, &adap->mps_ref, list) {
 | |
| 		list_del(&mps_entry->list);
 | |
| 		kfree(mps_entry);
 | |
| 	}
 | |
| 	spin_unlock(&adap->mps_ref_lock);
 | |
| }
 |