You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
Merge branch 'master' of master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6
Conflicts: drivers/net/yellowfin.c
This commit is contained in:
+3
-1
@@ -1727,12 +1727,14 @@ config KS8842
|
||||
tristate "Micrel KSZ8842"
|
||||
depends on HAS_IOMEM
|
||||
help
|
||||
This platform driver is for Micrel KSZ8842 chip.
|
||||
This platform driver is for Micrel KSZ8842 / KS8842
|
||||
2-port ethernet switch chip (managed, VLAN, QoS).
|
||||
|
||||
config KS8851
|
||||
tristate "Micrel KS8851 SPI"
|
||||
depends on SPI
|
||||
select MII
|
||||
select CRC32
|
||||
help
|
||||
SPI driver for Micrel KS8851 SPI attached network chip.
|
||||
|
||||
|
||||
@@ -1098,7 +1098,7 @@ static struct platform_driver w90p910_ether_driver = {
|
||||
.probe = w90p910_ether_probe,
|
||||
.remove = __devexit_p(w90p910_ether_remove),
|
||||
.driver = {
|
||||
.name = "w90p910-emc",
|
||||
.name = "nuc900-emc",
|
||||
.owner = THIS_MODULE,
|
||||
},
|
||||
};
|
||||
@@ -1119,5 +1119,5 @@ module_exit(w90p910_ether_exit);
|
||||
MODULE_AUTHOR("Wan ZongShun <mcuos.com@gmail.com>");
|
||||
MODULE_DESCRIPTION("w90p910 MAC driver!");
|
||||
MODULE_LICENSE("GPL");
|
||||
MODULE_ALIAS("platform:w90p910-emc");
|
||||
MODULE_ALIAS("platform:nuc900-emc");
|
||||
|
||||
|
||||
+3
-2
@@ -952,9 +952,10 @@ static netdev_tx_t b44_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
int rc = NETDEV_TX_OK;
|
||||
dma_addr_t mapping;
|
||||
u32 len, entry, ctrl;
|
||||
unsigned long flags;
|
||||
|
||||
len = skb->len;
|
||||
spin_lock_irq(&bp->lock);
|
||||
spin_lock_irqsave(&bp->lock, flags);
|
||||
|
||||
/* This is a hard error, log it. */
|
||||
if (unlikely(TX_BUFFS_AVAIL(bp) < 1)) {
|
||||
@@ -1027,7 +1028,7 @@ static netdev_tx_t b44_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
dev->trans_start = jiffies;
|
||||
|
||||
out_unlock:
|
||||
spin_unlock_irq(&bp->lock);
|
||||
spin_unlock_irqrestore(&bp->lock, flags);
|
||||
|
||||
return rc;
|
||||
|
||||
|
||||
+11
-6
@@ -401,9 +401,11 @@ static int bnx2_unregister_cnic(struct net_device *dev)
|
||||
struct bnx2_napi *bnapi = &bp->bnx2_napi[0];
|
||||
struct cnic_eth_dev *cp = &bp->cnic_eth_dev;
|
||||
|
||||
mutex_lock(&bp->cnic_lock);
|
||||
cp->drv_state = 0;
|
||||
bnapi->cnic_present = 0;
|
||||
rcu_assign_pointer(bp->cnic_ops, NULL);
|
||||
mutex_unlock(&bp->cnic_lock);
|
||||
synchronize_rcu();
|
||||
return 0;
|
||||
}
|
||||
@@ -431,13 +433,13 @@ bnx2_cnic_stop(struct bnx2 *bp)
|
||||
struct cnic_ops *c_ops;
|
||||
struct cnic_ctl_info info;
|
||||
|
||||
rcu_read_lock();
|
||||
c_ops = rcu_dereference(bp->cnic_ops);
|
||||
mutex_lock(&bp->cnic_lock);
|
||||
c_ops = bp->cnic_ops;
|
||||
if (c_ops) {
|
||||
info.cmd = CNIC_CTL_STOP_CMD;
|
||||
c_ops->cnic_ctl(bp->cnic_data, &info);
|
||||
}
|
||||
rcu_read_unlock();
|
||||
mutex_unlock(&bp->cnic_lock);
|
||||
}
|
||||
|
||||
static void
|
||||
@@ -446,8 +448,8 @@ bnx2_cnic_start(struct bnx2 *bp)
|
||||
struct cnic_ops *c_ops;
|
||||
struct cnic_ctl_info info;
|
||||
|
||||
rcu_read_lock();
|
||||
c_ops = rcu_dereference(bp->cnic_ops);
|
||||
mutex_lock(&bp->cnic_lock);
|
||||
c_ops = bp->cnic_ops;
|
||||
if (c_ops) {
|
||||
if (!(bp->flags & BNX2_FLAG_USING_MSIX)) {
|
||||
struct bnx2_napi *bnapi = &bp->bnx2_napi[0];
|
||||
@@ -457,7 +459,7 @@ bnx2_cnic_start(struct bnx2 *bp)
|
||||
info.cmd = CNIC_CTL_START_CMD;
|
||||
c_ops->cnic_ctl(bp->cnic_data, &info);
|
||||
}
|
||||
rcu_read_unlock();
|
||||
mutex_unlock(&bp->cnic_lock);
|
||||
}
|
||||
|
||||
#else
|
||||
@@ -7687,6 +7689,9 @@ bnx2_init_board(struct pci_dev *pdev, struct net_device *dev)
|
||||
|
||||
spin_lock_init(&bp->phy_lock);
|
||||
spin_lock_init(&bp->indirect_lock);
|
||||
#ifdef BCM_CNIC
|
||||
mutex_init(&bp->cnic_lock);
|
||||
#endif
|
||||
INIT_WORK(&bp->reset_task, bnx2_reset_task);
|
||||
|
||||
dev->base_addr = dev->mem_start = pci_resource_start(pdev, 0);
|
||||
|
||||
@@ -6903,6 +6903,7 @@ struct bnx2 {
|
||||
u32 idle_chk_status_idx;
|
||||
|
||||
#ifdef BCM_CNIC
|
||||
struct mutex cnic_lock;
|
||||
struct cnic_eth_dev cnic_eth_dev;
|
||||
#endif
|
||||
|
||||
|
||||
+105
-40
@@ -138,6 +138,16 @@ static struct cnic_dev *cnic_from_netdev(struct net_device *netdev)
|
||||
return NULL;
|
||||
}
|
||||
|
||||
static inline void ulp_get(struct cnic_ulp_ops *ulp_ops)
|
||||
{
|
||||
atomic_inc(&ulp_ops->ref_count);
|
||||
}
|
||||
|
||||
static inline void ulp_put(struct cnic_ulp_ops *ulp_ops)
|
||||
{
|
||||
atomic_dec(&ulp_ops->ref_count);
|
||||
}
|
||||
|
||||
static void cnic_ctx_wr(struct cnic_dev *dev, u32 cid_addr, u32 off, u32 val)
|
||||
{
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
@@ -358,6 +368,7 @@ int cnic_register_driver(int ulp_type, struct cnic_ulp_ops *ulp_ops)
|
||||
}
|
||||
read_unlock(&cnic_dev_lock);
|
||||
|
||||
atomic_set(&ulp_ops->ref_count, 0);
|
||||
rcu_assign_pointer(cnic_ulp_tbl[ulp_type], ulp_ops);
|
||||
mutex_unlock(&cnic_lock);
|
||||
|
||||
@@ -379,6 +390,8 @@ int cnic_register_driver(int ulp_type, struct cnic_ulp_ops *ulp_ops)
|
||||
int cnic_unregister_driver(int ulp_type)
|
||||
{
|
||||
struct cnic_dev *dev;
|
||||
struct cnic_ulp_ops *ulp_ops;
|
||||
int i = 0;
|
||||
|
||||
if (ulp_type >= MAX_CNIC_ULP_TYPE) {
|
||||
printk(KERN_ERR PFX "cnic_unregister_driver: Bad type %d\n",
|
||||
@@ -386,7 +399,8 @@ int cnic_unregister_driver(int ulp_type)
|
||||
return -EINVAL;
|
||||
}
|
||||
mutex_lock(&cnic_lock);
|
||||
if (!cnic_ulp_tbl[ulp_type]) {
|
||||
ulp_ops = cnic_ulp_tbl[ulp_type];
|
||||
if (!ulp_ops) {
|
||||
printk(KERN_ERR PFX "cnic_unregister_driver: Type %d has not "
|
||||
"been registered\n", ulp_type);
|
||||
goto out_unlock;
|
||||
@@ -411,6 +425,14 @@ int cnic_unregister_driver(int ulp_type)
|
||||
|
||||
mutex_unlock(&cnic_lock);
|
||||
synchronize_rcu();
|
||||
while ((atomic_read(&ulp_ops->ref_count) != 0) && (i < 20)) {
|
||||
msleep(100);
|
||||
i++;
|
||||
}
|
||||
|
||||
if (atomic_read(&ulp_ops->ref_count) != 0)
|
||||
printk(KERN_WARNING PFX "%s: Failed waiting for ref count to go"
|
||||
" to zero.\n", dev->netdev->name);
|
||||
return 0;
|
||||
|
||||
out_unlock:
|
||||
@@ -466,6 +488,7 @@ EXPORT_SYMBOL(cnic_register_driver);
|
||||
static int cnic_unregister_device(struct cnic_dev *dev, int ulp_type)
|
||||
{
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
int i = 0;
|
||||
|
||||
if (ulp_type >= MAX_CNIC_ULP_TYPE) {
|
||||
printk(KERN_ERR PFX "cnic_unregister_device: Bad type %d\n",
|
||||
@@ -486,6 +509,15 @@ static int cnic_unregister_device(struct cnic_dev *dev, int ulp_type)
|
||||
|
||||
synchronize_rcu();
|
||||
|
||||
while (test_bit(ULP_F_CALL_PENDING, &cp->ulp_flags[ulp_type]) &&
|
||||
i < 20) {
|
||||
msleep(100);
|
||||
i++;
|
||||
}
|
||||
if (test_bit(ULP_F_CALL_PENDING, &cp->ulp_flags[ulp_type]))
|
||||
printk(KERN_WARNING PFX "%s: Failed waiting for ULP up call"
|
||||
" to complete.\n", dev->netdev->name);
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(cnic_unregister_driver);
|
||||
@@ -1101,18 +1133,23 @@ static void cnic_ulp_stop(struct cnic_dev *dev)
|
||||
if (cp->cnic_uinfo)
|
||||
cnic_send_nlmsg(cp, ISCSI_KEVENT_IF_DOWN, NULL);
|
||||
|
||||
rcu_read_lock();
|
||||
for (if_type = 0; if_type < MAX_CNIC_ULP_TYPE; if_type++) {
|
||||
struct cnic_ulp_ops *ulp_ops;
|
||||
|
||||
ulp_ops = rcu_dereference(cp->ulp_ops[if_type]);
|
||||
if (!ulp_ops)
|
||||
mutex_lock(&cnic_lock);
|
||||
ulp_ops = cp->ulp_ops[if_type];
|
||||
if (!ulp_ops) {
|
||||
mutex_unlock(&cnic_lock);
|
||||
continue;
|
||||
}
|
||||
set_bit(ULP_F_CALL_PENDING, &cp->ulp_flags[if_type]);
|
||||
mutex_unlock(&cnic_lock);
|
||||
|
||||
if (test_and_clear_bit(ULP_F_START, &cp->ulp_flags[if_type]))
|
||||
ulp_ops->cnic_stop(cp->ulp_handle[if_type]);
|
||||
|
||||
clear_bit(ULP_F_CALL_PENDING, &cp->ulp_flags[if_type]);
|
||||
}
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
static void cnic_ulp_start(struct cnic_dev *dev)
|
||||
@@ -1120,18 +1157,23 @@ static void cnic_ulp_start(struct cnic_dev *dev)
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
int if_type;
|
||||
|
||||
rcu_read_lock();
|
||||
for (if_type = 0; if_type < MAX_CNIC_ULP_TYPE; if_type++) {
|
||||
struct cnic_ulp_ops *ulp_ops;
|
||||
|
||||
ulp_ops = rcu_dereference(cp->ulp_ops[if_type]);
|
||||
if (!ulp_ops || !ulp_ops->cnic_start)
|
||||
mutex_lock(&cnic_lock);
|
||||
ulp_ops = cp->ulp_ops[if_type];
|
||||
if (!ulp_ops || !ulp_ops->cnic_start) {
|
||||
mutex_unlock(&cnic_lock);
|
||||
continue;
|
||||
}
|
||||
set_bit(ULP_F_CALL_PENDING, &cp->ulp_flags[if_type]);
|
||||
mutex_unlock(&cnic_lock);
|
||||
|
||||
if (!test_and_set_bit(ULP_F_START, &cp->ulp_flags[if_type]))
|
||||
ulp_ops->cnic_start(cp->ulp_handle[if_type]);
|
||||
|
||||
clear_bit(ULP_F_CALL_PENDING, &cp->ulp_flags[if_type]);
|
||||
}
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
static int cnic_ctl(void *data, struct cnic_ctl_info *info)
|
||||
@@ -1141,22 +1183,18 @@ static int cnic_ctl(void *data, struct cnic_ctl_info *info)
|
||||
switch (info->cmd) {
|
||||
case CNIC_CTL_STOP_CMD:
|
||||
cnic_hold(dev);
|
||||
mutex_lock(&cnic_lock);
|
||||
|
||||
cnic_ulp_stop(dev);
|
||||
cnic_stop_hw(dev);
|
||||
|
||||
mutex_unlock(&cnic_lock);
|
||||
cnic_put(dev);
|
||||
break;
|
||||
case CNIC_CTL_START_CMD:
|
||||
cnic_hold(dev);
|
||||
mutex_lock(&cnic_lock);
|
||||
|
||||
if (!cnic_start_hw(dev))
|
||||
cnic_ulp_start(dev);
|
||||
|
||||
mutex_unlock(&cnic_lock);
|
||||
cnic_put(dev);
|
||||
break;
|
||||
default:
|
||||
@@ -1170,19 +1208,23 @@ static void cnic_ulp_init(struct cnic_dev *dev)
|
||||
int i;
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
|
||||
rcu_read_lock();
|
||||
for (i = 0; i < MAX_CNIC_ULP_TYPE_EXT; i++) {
|
||||
struct cnic_ulp_ops *ulp_ops;
|
||||
|
||||
ulp_ops = rcu_dereference(cnic_ulp_tbl[i]);
|
||||
if (!ulp_ops || !ulp_ops->cnic_init)
|
||||
mutex_lock(&cnic_lock);
|
||||
ulp_ops = cnic_ulp_tbl[i];
|
||||
if (!ulp_ops || !ulp_ops->cnic_init) {
|
||||
mutex_unlock(&cnic_lock);
|
||||
continue;
|
||||
}
|
||||
ulp_get(ulp_ops);
|
||||
mutex_unlock(&cnic_lock);
|
||||
|
||||
if (!test_and_set_bit(ULP_F_INIT, &cp->ulp_flags[i]))
|
||||
ulp_ops->cnic_init(dev);
|
||||
|
||||
ulp_put(ulp_ops);
|
||||
}
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
static void cnic_ulp_exit(struct cnic_dev *dev)
|
||||
@@ -1190,19 +1232,23 @@ static void cnic_ulp_exit(struct cnic_dev *dev)
|
||||
int i;
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
|
||||
rcu_read_lock();
|
||||
for (i = 0; i < MAX_CNIC_ULP_TYPE_EXT; i++) {
|
||||
struct cnic_ulp_ops *ulp_ops;
|
||||
|
||||
ulp_ops = rcu_dereference(cnic_ulp_tbl[i]);
|
||||
if (!ulp_ops || !ulp_ops->cnic_exit)
|
||||
mutex_lock(&cnic_lock);
|
||||
ulp_ops = cnic_ulp_tbl[i];
|
||||
if (!ulp_ops || !ulp_ops->cnic_exit) {
|
||||
mutex_unlock(&cnic_lock);
|
||||
continue;
|
||||
}
|
||||
ulp_get(ulp_ops);
|
||||
mutex_unlock(&cnic_lock);
|
||||
|
||||
if (test_and_clear_bit(ULP_F_INIT, &cp->ulp_flags[i]))
|
||||
ulp_ops->cnic_exit(dev);
|
||||
|
||||
ulp_put(ulp_ops);
|
||||
}
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
static int cnic_cm_offload_pg(struct cnic_sock *csk)
|
||||
@@ -2418,6 +2464,37 @@ static int cnic_start_bnx2_hw(struct cnic_dev *dev)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int cnic_register_netdev(struct cnic_dev *dev)
|
||||
{
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
struct cnic_eth_dev *ethdev = cp->ethdev;
|
||||
int err;
|
||||
|
||||
if (!ethdev)
|
||||
return -ENODEV;
|
||||
|
||||
if (ethdev->drv_state & CNIC_DRV_STATE_REGD)
|
||||
return 0;
|
||||
|
||||
err = ethdev->drv_register_cnic(dev->netdev, cp->cnic_ops, dev);
|
||||
if (err)
|
||||
printk(KERN_ERR PFX "%s: register_cnic failed\n",
|
||||
dev->netdev->name);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
static void cnic_unregister_netdev(struct cnic_dev *dev)
|
||||
{
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
struct cnic_eth_dev *ethdev = cp->ethdev;
|
||||
|
||||
if (!ethdev)
|
||||
return;
|
||||
|
||||
ethdev->drv_unregister_cnic(dev->netdev);
|
||||
}
|
||||
|
||||
static int cnic_start_hw(struct cnic_dev *dev)
|
||||
{
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
@@ -2427,13 +2504,6 @@ static int cnic_start_hw(struct cnic_dev *dev)
|
||||
if (test_bit(CNIC_F_CNIC_UP, &dev->flags))
|
||||
return -EALREADY;
|
||||
|
||||
err = ethdev->drv_register_cnic(dev->netdev, cp->cnic_ops, dev);
|
||||
if (err) {
|
||||
printk(KERN_ERR PFX "%s: register_cnic failed\n",
|
||||
dev->netdev->name);
|
||||
goto err2;
|
||||
}
|
||||
|
||||
dev->regview = ethdev->io_base;
|
||||
cp->chip_id = ethdev->chip_id;
|
||||
pci_dev_get(dev->pcidev);
|
||||
@@ -2463,18 +2533,13 @@ static int cnic_start_hw(struct cnic_dev *dev)
|
||||
return 0;
|
||||
|
||||
err1:
|
||||
ethdev->drv_unregister_cnic(dev->netdev);
|
||||
cp->free_resc(dev);
|
||||
pci_dev_put(dev->pcidev);
|
||||
err2:
|
||||
return err;
|
||||
}
|
||||
|
||||
static void cnic_stop_bnx2_hw(struct cnic_dev *dev)
|
||||
{
|
||||
struct cnic_local *cp = dev->cnic_priv;
|
||||
struct cnic_eth_dev *ethdev = cp->ethdev;
|
||||
|
||||
cnic_disable_bnx2_int_sync(dev);
|
||||
|
||||
cnic_reg_wr_ind(dev, BNX2_CP_SCRATCH + 0x20, 0);
|
||||
@@ -2486,8 +2551,6 @@ static void cnic_stop_bnx2_hw(struct cnic_dev *dev)
|
||||
cnic_setup_5709_context(dev, 0);
|
||||
cnic_free_irq(dev);
|
||||
|
||||
ethdev->drv_unregister_cnic(dev->netdev);
|
||||
|
||||
cnic_free_resc(dev);
|
||||
}
|
||||
|
||||
@@ -2568,7 +2631,7 @@ static struct cnic_dev *init_bnx2_cnic(struct net_device *dev)
|
||||
probe = symbol_get(bnx2_cnic_probe);
|
||||
if (probe) {
|
||||
ethdev = (*probe)(dev);
|
||||
symbol_put_addr(probe);
|
||||
symbol_put(bnx2_cnic_probe);
|
||||
}
|
||||
if (!ethdev)
|
||||
return NULL;
|
||||
@@ -2671,10 +2734,12 @@ static int cnic_netdev_event(struct notifier_block *this, unsigned long event,
|
||||
else if (event == NETDEV_UNREGISTER)
|
||||
cnic_ulp_exit(dev);
|
||||
else if (event == NETDEV_UP) {
|
||||
mutex_lock(&cnic_lock);
|
||||
if (cnic_register_netdev(dev) != 0) {
|
||||
cnic_put(dev);
|
||||
goto done;
|
||||
}
|
||||
if (!cnic_start_hw(dev))
|
||||
cnic_ulp_start(dev);
|
||||
mutex_unlock(&cnic_lock);
|
||||
}
|
||||
|
||||
rcu_read_lock();
|
||||
@@ -2693,10 +2758,9 @@ static int cnic_netdev_event(struct notifier_block *this, unsigned long event,
|
||||
rcu_read_unlock();
|
||||
|
||||
if (event == NETDEV_GOING_DOWN) {
|
||||
mutex_lock(&cnic_lock);
|
||||
cnic_ulp_stop(dev);
|
||||
cnic_stop_hw(dev);
|
||||
mutex_unlock(&cnic_lock);
|
||||
cnic_unregister_netdev(dev);
|
||||
} else if (event == NETDEV_UNREGISTER) {
|
||||
write_lock(&cnic_dev_lock);
|
||||
list_del_init(&dev->list);
|
||||
@@ -2728,6 +2792,7 @@ static void cnic_release(void)
|
||||
}
|
||||
|
||||
cnic_ulp_exit(dev);
|
||||
cnic_unregister_netdev(dev);
|
||||
list_del_init(&dev->list);
|
||||
cnic_free_dev(dev);
|
||||
}
|
||||
|
||||
@@ -176,6 +176,7 @@ struct cnic_local {
|
||||
unsigned long ulp_flags[MAX_CNIC_ULP_TYPE];
|
||||
#define ULP_F_INIT 0
|
||||
#define ULP_F_START 1
|
||||
#define ULP_F_CALL_PENDING 2
|
||||
struct cnic_ulp_ops *ulp_ops[MAX_CNIC_ULP_TYPE];
|
||||
|
||||
/* protected by ulp_lock */
|
||||
|
||||
@@ -290,6 +290,7 @@ struct cnic_ulp_ops {
|
||||
void (*iscsi_nl_send_msg)(struct cnic_dev *dev, u32 msg_type,
|
||||
char *data, u16 data_size);
|
||||
struct module *owner;
|
||||
atomic_t ref_count;
|
||||
};
|
||||
|
||||
extern int cnic_register_driver(int ulp_type, struct cnic_ulp_ops *ulp_ops);
|
||||
|
||||
+1
-1
@@ -1900,7 +1900,7 @@ static int e100_rx_indicate(struct nic *nic, struct rx *rx,
|
||||
nic->ru_running = RU_SUSPENDED;
|
||||
pci_dma_sync_single_for_device(nic->pdev, rx->dma_addr,
|
||||
sizeof(struct rfd),
|
||||
PCI_DMA_BIDIRECTIONAL);
|
||||
PCI_DMA_FROMDEVICE);
|
||||
return -ENODATA;
|
||||
}
|
||||
|
||||
|
||||
+11
-11
@@ -4539,8 +4539,7 @@ static int __e1000_shutdown(struct pci_dev *pdev, bool *enable_wake)
|
||||
/* Allow time for pending master requests to run */
|
||||
e1000e_disable_pcie_master(&adapter->hw);
|
||||
|
||||
if ((adapter->flags2 & FLAG2_HAS_PHY_WAKEUP) &&
|
||||
!(hw->mac.ops.check_mng_mode(hw))) {
|
||||
if (adapter->flags2 & FLAG2_HAS_PHY_WAKEUP) {
|
||||
/* enable wakeup by the PHY */
|
||||
retval = e1000_init_phy_wakeup(adapter, wufc);
|
||||
if (retval)
|
||||
@@ -4558,7 +4557,8 @@ static int __e1000_shutdown(struct pci_dev *pdev, bool *enable_wake)
|
||||
*enable_wake = !!wufc;
|
||||
|
||||
/* make sure adapter isn't asleep if manageability is enabled */
|
||||
if (adapter->flags & FLAG_MNG_PT_ENABLED)
|
||||
if ((adapter->flags & FLAG_MNG_PT_ENABLED) ||
|
||||
(hw->mac.ops.check_mng_mode(hw)))
|
||||
*enable_wake = true;
|
||||
|
||||
if (adapter->hw.phy.type == e1000_phy_igp_3)
|
||||
@@ -4671,14 +4671,6 @@ static int e1000_resume(struct pci_dev *pdev)
|
||||
return err;
|
||||
}
|
||||
|
||||
/* AER (Advanced Error Reporting) hooks */
|
||||
err = pci_enable_pcie_error_reporting(pdev);
|
||||
if (err) {
|
||||
dev_err(&pdev->dev, "pci_enable_pcie_error_reporting failed "
|
||||
"0x%x\n", err);
|
||||
/* non-fatal, continue */
|
||||
}
|
||||
|
||||
pci_set_master(pdev);
|
||||
|
||||
pci_enable_wake(pdev, PCI_D3hot, 0);
|
||||
@@ -4991,6 +4983,14 @@ static int __devinit e1000_probe(struct pci_dev *pdev,
|
||||
if (err)
|
||||
goto err_pci_reg;
|
||||
|
||||
/* AER (Advanced Error Reporting) hooks */
|
||||
err = pci_enable_pcie_error_reporting(pdev);
|
||||
if (err) {
|
||||
dev_err(&pdev->dev, "pci_enable_pcie_error_reporting failed "
|
||||
"0x%x\n", err);
|
||||
/* non-fatal, continue */
|
||||
}
|
||||
|
||||
pci_set_master(pdev);
|
||||
/* PCI config space info */
|
||||
err = pci_save_state(pdev);
|
||||
|
||||
@@ -309,6 +309,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
{
|
||||
struct mpc52xx_fec_priv *priv = netdev_priv(dev);
|
||||
struct bcom_fec_bd *bd;
|
||||
unsigned long flags;
|
||||
|
||||
if (bcom_queue_full(priv->tx_dmatsk)) {
|
||||
if (net_ratelimit())
|
||||
@@ -316,7 +317,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
return NETDEV_TX_BUSY;
|
||||
}
|
||||
|
||||
spin_lock_irq(&priv->lock);
|
||||
spin_lock_irqsave(&priv->lock, flags);
|
||||
dev->trans_start = jiffies;
|
||||
|
||||
bd = (struct bcom_fec_bd *)
|
||||
@@ -332,7 +333,7 @@ static int mpc52xx_fec_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
netif_stop_queue(dev);
|
||||
}
|
||||
|
||||
spin_unlock_irq(&priv->lock);
|
||||
spin_unlock_irqrestore(&priv->lock, flags);
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
|
||||
@@ -490,6 +490,7 @@ static int gfar_remove(struct of_device *ofdev)
|
||||
|
||||
dev_set_drvdata(&ofdev->dev, NULL);
|
||||
|
||||
unregister_netdev(dev);
|
||||
iounmap(priv->regs);
|
||||
free_netdev(priv->ndev);
|
||||
|
||||
|
||||
@@ -1305,6 +1305,8 @@ static int emac_close(struct net_device *ndev)
|
||||
|
||||
free_irq(dev->emac_irq, dev);
|
||||
|
||||
netif_carrier_off(ndev);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -23,7 +23,6 @@
|
||||
#include <linux/init.h>
|
||||
#include <linux/errno.h>
|
||||
#include <linux/netdevice.h>
|
||||
#include <linux/etherdevice.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/rtnetlink.h>
|
||||
#include <linux/interrupt.h>
|
||||
@@ -205,9 +204,6 @@ static const struct net_device_ops au1k_irda_netdev_ops = {
|
||||
.ndo_start_xmit = au1k_irda_hard_xmit,
|
||||
.ndo_tx_timeout = au1k_tx_timeout,
|
||||
.ndo_do_ioctl = au1k_irda_ioctl,
|
||||
.ndo_change_mtu = eth_change_mtu,
|
||||
.ndo_validate_addr = eth_validate_addr,
|
||||
.ndo_set_mac_address = eth_mac_addr,
|
||||
};
|
||||
|
||||
static int au1k_irda_net_init(struct net_device *dev)
|
||||
|
||||
@@ -803,9 +803,6 @@ static const struct net_device_ops pxa_irda_netdev_ops = {
|
||||
.ndo_stop = pxa_irda_stop,
|
||||
.ndo_start_xmit = pxa_irda_hard_xmit,
|
||||
.ndo_do_ioctl = pxa_irda_ioctl,
|
||||
.ndo_change_mtu = eth_change_mtu,
|
||||
.ndo_validate_addr = eth_validate_addr,
|
||||
.ndo_set_mac_address = eth_mac_addr,
|
||||
};
|
||||
|
||||
static int pxa_irda_probe(struct platform_device *pdev)
|
||||
@@ -830,6 +827,7 @@ static int pxa_irda_probe(struct platform_device *pdev)
|
||||
if (!dev)
|
||||
goto err_mem_3;
|
||||
|
||||
SET_NETDEV_DEV(dev, &pdev->dev);
|
||||
si = netdev_priv(dev);
|
||||
si->dev = &pdev->dev;
|
||||
si->pdata = pdev->dev.platform_data;
|
||||
|
||||
@@ -24,7 +24,6 @@
|
||||
#include <linux/init.h>
|
||||
#include <linux/errno.h>
|
||||
#include <linux/netdevice.h>
|
||||
#include <linux/etherdevice.h>
|
||||
#include <linux/slab.h>
|
||||
#include <linux/rtnetlink.h>
|
||||
#include <linux/interrupt.h>
|
||||
@@ -881,9 +880,6 @@ static const struct net_device_ops sa1100_irda_netdev_ops = {
|
||||
.ndo_stop = sa1100_irda_stop,
|
||||
.ndo_start_xmit = sa1100_irda_hard_xmit,
|
||||
.ndo_do_ioctl = sa1100_irda_ioctl,
|
||||
.ndo_change_mtu = eth_change_mtu,
|
||||
.ndo_validate_addr = eth_validate_addr,
|
||||
.ndo_set_mac_address = eth_mac_addr,
|
||||
};
|
||||
|
||||
static int sa1100_irda_probe(struct platform_device *pdev)
|
||||
|
||||
@@ -41,6 +41,7 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
struct ixpdev_priv *ip = netdev_priv(dev);
|
||||
struct ixpdev_tx_desc *desc;
|
||||
int entry;
|
||||
unsigned long flags;
|
||||
|
||||
if (unlikely(skb->len > PAGE_SIZE)) {
|
||||
/* @@@ Count drops. */
|
||||
@@ -63,11 +64,11 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
|
||||
dev->trans_start = jiffies;
|
||||
|
||||
local_irq_disable();
|
||||
local_irq_save(flags);
|
||||
ip->tx_queue_entries++;
|
||||
if (ip->tx_queue_entries == TX_BUF_COUNT_PER_CHAN)
|
||||
netif_stop_queue(dev);
|
||||
local_irq_enable();
|
||||
local_irq_restore(flags);
|
||||
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
|
||||
+4
-3
@@ -620,6 +620,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
dma_addr_t mapping;
|
||||
unsigned int len, entry;
|
||||
u32 ctrl;
|
||||
unsigned long flags;
|
||||
|
||||
#ifdef DEBUG
|
||||
int i;
|
||||
@@ -635,12 +636,12 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
#endif
|
||||
|
||||
len = skb->len;
|
||||
spin_lock_irq(&bp->lock);
|
||||
spin_lock_irqsave(&bp->lock, flags);
|
||||
|
||||
/* This is a hard error, log it. */
|
||||
if (TX_BUFFS_AVAIL(bp) < 1) {
|
||||
netif_stop_queue(dev);
|
||||
spin_unlock_irq(&bp->lock);
|
||||
spin_unlock_irqrestore(&bp->lock, flags);
|
||||
dev_err(&bp->pdev->dev,
|
||||
"BUG! Tx Ring full when queue awake!\n");
|
||||
dev_dbg(&bp->pdev->dev, "tx_head = %u, tx_tail = %u\n",
|
||||
@@ -674,7 +675,7 @@ static int macb_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
if (TX_BUFFS_AVAIL(bp) < 1)
|
||||
netif_stop_queue(dev);
|
||||
|
||||
spin_unlock_irq(&bp->lock);
|
||||
spin_unlock_irqrestore(&bp->lock, flags);
|
||||
|
||||
dev->trans_start = jiffies;
|
||||
|
||||
|
||||
@@ -437,6 +437,7 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind)
|
||||
{
|
||||
struct mlx4_en_cq *cq = &priv->tx_cq[tx_ind];
|
||||
struct mlx4_en_tx_ring *ring = &priv->tx_ring[tx_ind];
|
||||
unsigned long flags;
|
||||
|
||||
/* If we don't have a pending timer, set one up to catch our recent
|
||||
post in case the interface becomes idle */
|
||||
@@ -445,9 +446,9 @@ static inline void mlx4_en_xmit_poll(struct mlx4_en_priv *priv, int tx_ind)
|
||||
|
||||
/* Poll the CQ every mlx4_en_TX_MODER_POLL packets */
|
||||
if ((++ring->poll_cnt & (MLX4_EN_TX_POLL_MODER - 1)) == 0)
|
||||
if (spin_trylock_irq(&ring->comp_lock)) {
|
||||
if (spin_trylock_irqsave(&ring->comp_lock, flags)) {
|
||||
mlx4_en_process_tx_cq(priv->dev, cq);
|
||||
spin_unlock_irq(&ring->comp_lock);
|
||||
spin_unlock_irqrestore(&ring->comp_lock, flags);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
+22
-18
@@ -196,21 +196,23 @@ static void PRINT_PKT(u_char *buf, int length)
|
||||
/* this enables an interrupt in the interrupt mask register */
|
||||
#define SMC_ENABLE_INT(lp, x) do { \
|
||||
unsigned char mask; \
|
||||
spin_lock_irq(&lp->lock); \
|
||||
unsigned long smc_enable_flags; \
|
||||
spin_lock_irqsave(&lp->lock, smc_enable_flags); \
|
||||
mask = SMC_GET_INT_MASK(lp); \
|
||||
mask |= (x); \
|
||||
SMC_SET_INT_MASK(lp, mask); \
|
||||
spin_unlock_irq(&lp->lock); \
|
||||
spin_unlock_irqrestore(&lp->lock, smc_enable_flags); \
|
||||
} while (0)
|
||||
|
||||
/* this disables an interrupt from the interrupt mask register */
|
||||
#define SMC_DISABLE_INT(lp, x) do { \
|
||||
unsigned char mask; \
|
||||
spin_lock_irq(&lp->lock); \
|
||||
unsigned long smc_disable_flags; \
|
||||
spin_lock_irqsave(&lp->lock, smc_disable_flags); \
|
||||
mask = SMC_GET_INT_MASK(lp); \
|
||||
mask &= ~(x); \
|
||||
SMC_SET_INT_MASK(lp, mask); \
|
||||
spin_unlock_irq(&lp->lock); \
|
||||
spin_unlock_irqrestore(&lp->lock, smc_disable_flags); \
|
||||
} while (0)
|
||||
|
||||
/*
|
||||
@@ -520,21 +522,21 @@ static inline void smc_rcv(struct net_device *dev)
|
||||
* any other concurrent access and C would always interrupt B. But life
|
||||
* isn't that easy in a SMP world...
|
||||
*/
|
||||
#define smc_special_trylock(lock) \
|
||||
#define smc_special_trylock(lock, flags) \
|
||||
({ \
|
||||
int __ret; \
|
||||
local_irq_disable(); \
|
||||
local_irq_save(flags); \
|
||||
__ret = spin_trylock(lock); \
|
||||
if (!__ret) \
|
||||
local_irq_enable(); \
|
||||
local_irq_restore(flags); \
|
||||
__ret; \
|
||||
})
|
||||
#define smc_special_lock(lock) spin_lock_irq(lock)
|
||||
#define smc_special_unlock(lock) spin_unlock_irq(lock)
|
||||
#define smc_special_lock(lock, flags) spin_lock_irqsave(lock, flags)
|
||||
#define smc_special_unlock(lock, flags) spin_unlock_irqrestore(lock, flags)
|
||||
#else
|
||||
#define smc_special_trylock(lock) (1)
|
||||
#define smc_special_lock(lock) do { } while (0)
|
||||
#define smc_special_unlock(lock) do { } while (0)
|
||||
#define smc_special_trylock(lock, flags) (1)
|
||||
#define smc_special_lock(lock, flags) do { } while (0)
|
||||
#define smc_special_unlock(lock, flags) do { } while (0)
|
||||
#endif
|
||||
|
||||
/*
|
||||
@@ -548,10 +550,11 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||
struct sk_buff *skb;
|
||||
unsigned int packet_no, len;
|
||||
unsigned char *buf;
|
||||
unsigned long flags;
|
||||
|
||||
DBG(3, "%s: %s\n", dev->name, __func__);
|
||||
|
||||
if (!smc_special_trylock(&lp->lock)) {
|
||||
if (!smc_special_trylock(&lp->lock, flags)) {
|
||||
netif_stop_queue(dev);
|
||||
tasklet_schedule(&lp->tx_task);
|
||||
return;
|
||||
@@ -559,7 +562,7 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||
|
||||
skb = lp->pending_tx_skb;
|
||||
if (unlikely(!skb)) {
|
||||
smc_special_unlock(&lp->lock);
|
||||
smc_special_unlock(&lp->lock, flags);
|
||||
return;
|
||||
}
|
||||
lp->pending_tx_skb = NULL;
|
||||
@@ -569,7 +572,7 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||
printk("%s: Memory allocation failed.\n", dev->name);
|
||||
dev->stats.tx_errors++;
|
||||
dev->stats.tx_fifo_errors++;
|
||||
smc_special_unlock(&lp->lock);
|
||||
smc_special_unlock(&lp->lock, flags);
|
||||
goto done;
|
||||
}
|
||||
|
||||
@@ -608,7 +611,7 @@ static void smc_hardware_send_pkt(unsigned long data)
|
||||
|
||||
/* queue the packet for TX */
|
||||
SMC_SET_MMU_CMD(lp, MC_ENQUEUE);
|
||||
smc_special_unlock(&lp->lock);
|
||||
smc_special_unlock(&lp->lock, flags);
|
||||
|
||||
dev->trans_start = jiffies;
|
||||
dev->stats.tx_packets++;
|
||||
@@ -633,6 +636,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
struct smc_local *lp = netdev_priv(dev);
|
||||
void __iomem *ioaddr = lp->base;
|
||||
unsigned int numPages, poll_count, status;
|
||||
unsigned long flags;
|
||||
|
||||
DBG(3, "%s: %s\n", dev->name, __func__);
|
||||
|
||||
@@ -658,7 +662,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
return NETDEV_TX_OK;
|
||||
}
|
||||
|
||||
smc_special_lock(&lp->lock);
|
||||
smc_special_lock(&lp->lock, flags);
|
||||
|
||||
/* now, try to allocate the memory */
|
||||
SMC_SET_MMU_CMD(lp, MC_ALLOC | numPages);
|
||||
@@ -676,7 +680,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
|
||||
}
|
||||
} while (--poll_count);
|
||||
|
||||
smc_special_unlock(&lp->lock);
|
||||
smc_special_unlock(&lp->lock, flags);
|
||||
|
||||
lp->pending_tx_skb = skb;
|
||||
if (!poll_count) {
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user