You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
mellanox: Logging message cleanups
Use a more current logging style. o Coalesce formats o Add missing spaces for coalesced formats o Align arguments for modified formats o Add missing newlines for some logging messages o Use DRV_NAME as part of format instead of %s, DRV_NAME to reduce overall text. o Use ..., ##__VA_ARGS__ instead of args... in macros o Correct a few format typos o Use a single line message where appropriate Signed-off-by: Joe Perches <joe@perches.com> Acked-By: Amir Vadai <amirv@mellanox.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
committed by
David S. Miller
parent
2db2a15abf
commit
1a91de2883
@@ -212,8 +212,7 @@ static int mlx4_comm_cmd_poll(struct mlx4_dev *dev, u8 cmd, u16 param,
|
||||
|
||||
/* First, verify that the master reports correct status */
|
||||
if (comm_pending(dev)) {
|
||||
mlx4_warn(dev, "Communication channel is not idle."
|
||||
"my toggle is %d (cmd:0x%x)\n",
|
||||
mlx4_warn(dev, "Communication channel is not idle - my toggle is %d (cmd:0x%x)\n",
|
||||
priv->cmd.comm_toggle, cmd);
|
||||
return -EAGAIN;
|
||||
}
|
||||
@@ -422,9 +421,8 @@ static int mlx4_slave_cmd(struct mlx4_dev *dev, u64 in_param, u64 *out_param,
|
||||
*out_param =
|
||||
be64_to_cpu(vhcr->out_param);
|
||||
else {
|
||||
mlx4_err(dev, "response expected while"
|
||||
"output mailbox is NULL for "
|
||||
"command 0x%x\n", op);
|
||||
mlx4_err(dev, "response expected while output mailbox is NULL for command 0x%x\n",
|
||||
op);
|
||||
vhcr->status = CMD_STAT_BAD_PARAM;
|
||||
}
|
||||
}
|
||||
@@ -439,16 +437,15 @@ static int mlx4_slave_cmd(struct mlx4_dev *dev, u64 in_param, u64 *out_param,
|
||||
*out_param =
|
||||
be64_to_cpu(vhcr->out_param);
|
||||
else {
|
||||
mlx4_err(dev, "response expected while"
|
||||
"output mailbox is NULL for "
|
||||
"command 0x%x\n", op);
|
||||
mlx4_err(dev, "response expected while output mailbox is NULL for command 0x%x\n",
|
||||
op);
|
||||
vhcr->status = CMD_STAT_BAD_PARAM;
|
||||
}
|
||||
}
|
||||
ret = mlx4_status_to_errno(vhcr->status);
|
||||
} else
|
||||
mlx4_err(dev, "failed execution of VHCR_POST command"
|
||||
"opcode 0x%x\n", op);
|
||||
mlx4_err(dev, "failed execution of VHCR_POST command opcode 0x%x\n",
|
||||
op);
|
||||
}
|
||||
|
||||
mutex_unlock(&priv->cmd.slave_cmd_mutex);
|
||||
@@ -625,9 +622,8 @@ static int mlx4_ACCESS_MEM(struct mlx4_dev *dev, u64 master_addr,
|
||||
|
||||
if ((slave_addr & 0xfff) | (master_addr & 0xfff) |
|
||||
(slave & ~0x7f) | (size & 0xff)) {
|
||||
mlx4_err(dev, "Bad access mem params - slave_addr:0x%llx "
|
||||
"master_addr:0x%llx slave_id:%d size:%d\n",
|
||||
slave_addr, master_addr, slave, size);
|
||||
mlx4_err(dev, "Bad access mem params - slave_addr:0x%llx master_addr:0x%llx slave_id:%d size:%d\n",
|
||||
slave_addr, master_addr, slave, size);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@@ -788,8 +784,7 @@ static int mlx4_MAD_IFC_wrapper(struct mlx4_dev *dev, int slave,
|
||||
((smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) ||
|
||||
(smp->mgmt_class == IB_MGMT_CLASS_SUBN_LID_ROUTED &&
|
||||
smp->method == IB_MGMT_METHOD_SET))) {
|
||||
mlx4_err(dev, "slave %d is trying to execute a Subnet MGMT MAD, "
|
||||
"class 0x%x, method 0x%x for attr 0x%x. Rejecting\n",
|
||||
mlx4_err(dev, "slave %d is trying to execute a Subnet MGMT MAD, class 0x%x, method 0x%x for attr 0x%x - Rejecting\n",
|
||||
slave, smp->method, smp->mgmt_class,
|
||||
be16_to_cpu(smp->attr_id));
|
||||
return -EPERM;
|
||||
@@ -1409,8 +1404,8 @@ static int mlx4_master_process_vhcr(struct mlx4_dev *dev, int slave,
|
||||
ALIGN(sizeof(struct mlx4_vhcr_cmd),
|
||||
MLX4_ACCESS_MEM_ALIGN), 1);
|
||||
if (ret) {
|
||||
mlx4_err(dev, "%s:Failed reading vhcr"
|
||||
"ret: 0x%x\n", __func__, ret);
|
||||
mlx4_err(dev, "%s: Failed reading vhcr ret: 0x%x\n",
|
||||
__func__, ret);
|
||||
kfree(vhcr);
|
||||
return ret;
|
||||
}
|
||||
@@ -1461,9 +1456,8 @@ static int mlx4_master_process_vhcr(struct mlx4_dev *dev, int slave,
|
||||
|
||||
/* Apply permission and bound checks if applicable */
|
||||
if (cmd->verify && cmd->verify(dev, slave, vhcr, inbox)) {
|
||||
mlx4_warn(dev, "Command:0x%x from slave: %d failed protection "
|
||||
"checks for resource_id:%d\n", vhcr->op, slave,
|
||||
vhcr->in_modifier);
|
||||
mlx4_warn(dev, "Command:0x%x from slave: %d failed protection checks for resource_id:%d\n",
|
||||
vhcr->op, slave, vhcr->in_modifier);
|
||||
vhcr_cmd->status = CMD_STAT_BAD_OP;
|
||||
goto out_status;
|
||||
}
|
||||
@@ -1502,8 +1496,7 @@ static int mlx4_master_process_vhcr(struct mlx4_dev *dev, int slave,
|
||||
}
|
||||
|
||||
if (err) {
|
||||
mlx4_warn(dev, "vhcr command:0x%x slave:%d failed with"
|
||||
" error:%d, status %d\n",
|
||||
mlx4_warn(dev, "vhcr command:0x%x slave:%d failed with error:%d, status %d\n",
|
||||
vhcr->op, slave, vhcr->errno, err);
|
||||
vhcr_cmd->status = mlx4_errno_to_status(err);
|
||||
goto out_status;
|
||||
@@ -1537,8 +1530,8 @@ out_status:
|
||||
__func__);
|
||||
else if (vhcr->e_bit &&
|
||||
mlx4_GEN_EQE(dev, slave, &priv->mfunc.master.cmd_eqe))
|
||||
mlx4_warn(dev, "Failed to generate command completion "
|
||||
"eqe for slave %d\n", slave);
|
||||
mlx4_warn(dev, "Failed to generate command completion eqe for slave %d\n",
|
||||
slave);
|
||||
}
|
||||
|
||||
out:
|
||||
@@ -1577,8 +1570,9 @@ static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
|
||||
|
||||
mlx4_dbg(dev, "updating immediately admin params slave %d port %d\n",
|
||||
slave, port);
|
||||
mlx4_dbg(dev, "vlan %d QoS %d link down %d\n", vp_admin->default_vlan,
|
||||
vp_admin->default_qos, vp_admin->link_state);
|
||||
mlx4_dbg(dev, "vlan %d QoS %d link down %d\n",
|
||||
vp_admin->default_vlan, vp_admin->default_qos,
|
||||
vp_admin->link_state);
|
||||
|
||||
work = kzalloc(sizeof(*work), GFP_KERNEL);
|
||||
if (!work)
|
||||
@@ -1591,7 +1585,7 @@ static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
|
||||
&admin_vlan_ix);
|
||||
if (err) {
|
||||
kfree(work);
|
||||
mlx4_warn((&priv->dev),
|
||||
mlx4_warn(&priv->dev,
|
||||
"No vlan resources slave %d, port %d\n",
|
||||
slave, port);
|
||||
return err;
|
||||
@@ -1600,7 +1594,7 @@ static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
|
||||
admin_vlan_ix = NO_INDX;
|
||||
}
|
||||
work->flags |= MLX4_VF_IMMED_VLAN_FLAG_VLAN;
|
||||
mlx4_dbg((&(priv->dev)),
|
||||
mlx4_dbg(&priv->dev,
|
||||
"alloc vlan %d idx %d slave %d port %d\n",
|
||||
(int)(vp_admin->default_vlan),
|
||||
admin_vlan_ix, slave, port);
|
||||
@@ -1661,12 +1655,12 @@ static int mlx4_master_activate_admin_state(struct mlx4_priv *priv, int slave)
|
||||
vp_admin->default_vlan, &(vp_oper->vlan_idx));
|
||||
if (err) {
|
||||
vp_oper->vlan_idx = NO_INDX;
|
||||
mlx4_warn((&priv->dev),
|
||||
mlx4_warn(&priv->dev,
|
||||
"No vlan resorces slave %d, port %d\n",
|
||||
slave, port);
|
||||
return err;
|
||||
}
|
||||
mlx4_dbg((&(priv->dev)), "alloc vlan %d idx %d slave %d port %d\n",
|
||||
mlx4_dbg(&priv->dev, "alloc vlan %d idx %d slave %d port %d\n",
|
||||
(int)(vp_oper->state.default_vlan),
|
||||
vp_oper->vlan_idx, slave, port);
|
||||
}
|
||||
@@ -1677,12 +1671,12 @@ static int mlx4_master_activate_admin_state(struct mlx4_priv *priv, int slave)
|
||||
if (0 > vp_oper->mac_idx) {
|
||||
err = vp_oper->mac_idx;
|
||||
vp_oper->mac_idx = NO_INDX;
|
||||
mlx4_warn((&priv->dev),
|
||||
mlx4_warn(&priv->dev,
|
||||
"No mac resorces slave %d, port %d\n",
|
||||
slave, port);
|
||||
return err;
|
||||
}
|
||||
mlx4_dbg((&(priv->dev)), "alloc mac %llx idx %d slave %d port %d\n",
|
||||
mlx4_dbg(&priv->dev, "alloc mac %llx idx %d slave %d port %d\n",
|
||||
vp_oper->state.mac, vp_oper->mac_idx, slave, port);
|
||||
}
|
||||
}
|
||||
@@ -1731,8 +1725,8 @@ static void mlx4_master_do_cmd(struct mlx4_dev *dev, int slave, u8 cmd,
|
||||
slave_state[slave].comm_toggle ^= 1;
|
||||
reply = (u32) slave_state[slave].comm_toggle << 31;
|
||||
if (toggle != slave_state[slave].comm_toggle) {
|
||||
mlx4_warn(dev, "Incorrect toggle %d from slave %d. *** MASTER"
|
||||
"STATE COMPROMISIED ***\n", toggle, slave);
|
||||
mlx4_warn(dev, "Incorrect toggle %d from slave %d. *** MASTER STATE COMPROMISED ***\n",
|
||||
toggle, slave);
|
||||
goto reset_slave;
|
||||
}
|
||||
if (cmd == MLX4_COMM_CMD_RESET) {
|
||||
@@ -1759,8 +1753,8 @@ static void mlx4_master_do_cmd(struct mlx4_dev *dev, int slave, u8 cmd,
|
||||
/*command from slave in the middle of FLR*/
|
||||
if (cmd != MLX4_COMM_CMD_RESET &&
|
||||
MLX4_COMM_CMD_FLR == slave_state[slave].last_cmd) {
|
||||
mlx4_warn(dev, "slave:%d is Trying to run cmd(0x%x) "
|
||||
"in the middle of FLR\n", slave, cmd);
|
||||
mlx4_warn(dev, "slave:%d is Trying to run cmd(0x%x) in the middle of FLR\n",
|
||||
slave, cmd);
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -1798,8 +1792,8 @@ static void mlx4_master_do_cmd(struct mlx4_dev *dev, int slave, u8 cmd,
|
||||
|
||||
mutex_lock(&priv->cmd.slave_cmd_mutex);
|
||||
if (mlx4_master_process_vhcr(dev, slave, NULL)) {
|
||||
mlx4_err(dev, "Failed processing vhcr for slave:%d,"
|
||||
" resetting slave.\n", slave);
|
||||
mlx4_err(dev, "Failed processing vhcr for slave:%d, resetting slave\n",
|
||||
slave);
|
||||
mutex_unlock(&priv->cmd.slave_cmd_mutex);
|
||||
goto reset_slave;
|
||||
}
|
||||
@@ -1816,8 +1810,7 @@ static void mlx4_master_do_cmd(struct mlx4_dev *dev, int slave, u8 cmd,
|
||||
is_going_down = 1;
|
||||
spin_unlock_irqrestore(&priv->mfunc.master.slave_state_lock, flags);
|
||||
if (is_going_down) {
|
||||
mlx4_warn(dev, "Slave is going down aborting command(%d)"
|
||||
" executing from slave:%d\n",
|
||||
mlx4_warn(dev, "Slave is going down aborting command(%d) executing from slave:%d\n",
|
||||
cmd, slave);
|
||||
return;
|
||||
}
|
||||
@@ -1880,9 +1873,8 @@ void mlx4_master_comm_channel(struct work_struct *work)
|
||||
if (toggle != slt) {
|
||||
if (master->slave_state[slave].comm_toggle
|
||||
!= slt) {
|
||||
printk(KERN_INFO "slave %d out of sync."
|
||||
" read toggle %d, state toggle %d. "
|
||||
"Resynching.\n", slave, slt,
|
||||
printk(KERN_INFO "slave %d out of sync. read toggle %d, state toggle %d. Resynching.\n",
|
||||
slave, slt,
|
||||
master->slave_state[slave].comm_toggle);
|
||||
master->slave_state[slave].comm_toggle =
|
||||
slt;
|
||||
@@ -1896,8 +1888,7 @@ void mlx4_master_comm_channel(struct work_struct *work)
|
||||
}
|
||||
|
||||
if (reported && reported != served)
|
||||
mlx4_warn(dev, "Got command event with bitmask from %d slaves"
|
||||
" but %d were served\n",
|
||||
mlx4_warn(dev, "Got command event with bitmask from %d slaves but %d were served\n",
|
||||
reported, served);
|
||||
|
||||
if (mlx4_ARM_COMM_CHANNEL(dev))
|
||||
@@ -1953,7 +1944,7 @@ int mlx4_multi_func_init(struct mlx4_dev *dev)
|
||||
ioremap(pci_resource_start(dev->pdev, 2) +
|
||||
MLX4_SLAVE_COMM_BASE, MLX4_COMM_PAGESIZE);
|
||||
if (!priv->mfunc.comm) {
|
||||
mlx4_err(dev, "Couldn't map communication vector.\n");
|
||||
mlx4_err(dev, "Couldn't map communication vector\n");
|
||||
goto err_vhcr;
|
||||
}
|
||||
|
||||
@@ -2080,7 +2071,7 @@ int mlx4_cmd_init(struct mlx4_dev *dev)
|
||||
priv->cmd.hcr = ioremap(pci_resource_start(dev->pdev, 0) +
|
||||
MLX4_HCR_BASE, MLX4_HCR_SIZE);
|
||||
if (!priv->cmd.hcr) {
|
||||
mlx4_err(dev, "Couldn't map command register.\n");
|
||||
mlx4_err(dev, "Couldn't map command register\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -125,8 +125,7 @@ int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq,
|
||||
&cq->vector)) {
|
||||
cq->vector = (cq->ring + 1 + priv->port)
|
||||
% mdev->dev->caps.num_comp_vectors;
|
||||
mlx4_warn(mdev, "Failed Assigning an EQ to "
|
||||
"%s ,Falling back to legacy EQ's\n",
|
||||
mlx4_warn(mdev, "Failed assigning an EQ to %s, falling back to legacy EQ's\n",
|
||||
name);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -925,13 +925,13 @@ static int mlx4_en_flow_replace(struct net_device *dev,
|
||||
qpn = cmd->fs.ring_cookie & (EN_ETHTOOL_QP_ATTACH - 1);
|
||||
} else {
|
||||
if (cmd->fs.ring_cookie >= priv->rx_ring_num) {
|
||||
en_warn(priv, "rxnfc: RX ring (%llu) doesn't exist.\n",
|
||||
en_warn(priv, "rxnfc: RX ring (%llu) doesn't exist\n",
|
||||
cmd->fs.ring_cookie);
|
||||
return -EINVAL;
|
||||
}
|
||||
qpn = priv->rss_map.qps[cmd->fs.ring_cookie].qpn;
|
||||
if (!qpn) {
|
||||
en_warn(priv, "rxnfc: RX ring (%llu) is inactive.\n",
|
||||
en_warn(priv, "rxnfc: RX ring (%llu) is inactive\n",
|
||||
cmd->fs.ring_cookie);
|
||||
return -EINVAL;
|
||||
}
|
||||
@@ -956,7 +956,7 @@ static int mlx4_en_flow_replace(struct net_device *dev,
|
||||
}
|
||||
err = mlx4_flow_attach(priv->mdev->dev, &rule, ®_id);
|
||||
if (err) {
|
||||
en_err(priv, "Fail to attach network rule at location %d.\n",
|
||||
en_err(priv, "Fail to attach network rule at location %d\n",
|
||||
cmd->fs.location);
|
||||
goto out_free_list;
|
||||
}
|
||||
|
||||
@@ -133,7 +133,7 @@ static int mlx4_en_get_profile(struct mlx4_en_dev *mdev)
|
||||
MLX4_EN_MAX_TX_RING_P_UP);
|
||||
if (params->udp_rss && !(mdev->dev->caps.flags
|
||||
& MLX4_DEV_CAP_FLAG_UDP_RSS)) {
|
||||
mlx4_warn(mdev, "UDP RSS is not supported on this device.\n");
|
||||
mlx4_warn(mdev, "UDP RSS is not supported on this device\n");
|
||||
params->udp_rss = 0;
|
||||
}
|
||||
for (i = 1; i <= MLX4_MAX_PORTS; i++) {
|
||||
@@ -251,8 +251,7 @@ static void *mlx4_en_add(struct mlx4_dev *dev)
|
||||
|
||||
mdev->LSO_support = !!(dev->caps.flags & (1 << 15));
|
||||
if (!mdev->LSO_support)
|
||||
mlx4_warn(mdev, "LSO not supported, please upgrade to later "
|
||||
"FW version to enable LSO\n");
|
||||
mlx4_warn(mdev, "LSO not supported, please upgrade to later FW version to enable LSO\n");
|
||||
|
||||
if (mlx4_mr_alloc(mdev->dev, mdev->priv_pdn, 0, ~0ull,
|
||||
MLX4_PERM_LOCAL_WRITE | MLX4_PERM_LOCAL_READ,
|
||||
@@ -268,7 +267,7 @@ static void *mlx4_en_add(struct mlx4_dev *dev)
|
||||
/* Build device profile according to supplied module parameters */
|
||||
err = mlx4_en_get_profile(mdev);
|
||||
if (err) {
|
||||
mlx4_err(mdev, "Bad module parameters, aborting.\n");
|
||||
mlx4_err(mdev, "Bad module parameters, aborting\n");
|
||||
goto err_mr;
|
||||
}
|
||||
|
||||
|
||||
@@ -1576,7 +1576,7 @@ int mlx4_en_start_port(struct net_device *dev)
|
||||
cq->buf[j].owner_sr_opcode = MLX4_CQE_OWNER_MASK;
|
||||
err = mlx4_en_set_cq_moder(priv, cq);
|
||||
if (err) {
|
||||
en_err(priv, "Failed setting cq moderation parameters");
|
||||
en_err(priv, "Failed setting cq moderation parameters\n");
|
||||
mlx4_en_deactivate_cq(priv, cq);
|
||||
goto cq_err;
|
||||
}
|
||||
@@ -1615,7 +1615,7 @@ int mlx4_en_start_port(struct net_device *dev)
|
||||
}
|
||||
err = mlx4_en_set_cq_moder(priv, cq);
|
||||
if (err) {
|
||||
en_err(priv, "Failed setting cq moderation parameters");
|
||||
en_err(priv, "Failed setting cq moderation parameters\n");
|
||||
mlx4_en_deactivate_cq(priv, cq);
|
||||
goto tx_err;
|
||||
}
|
||||
@@ -2594,8 +2594,8 @@ int mlx4_en_init_netdev(struct mlx4_en_dev *mdev, int port,
|
||||
prof->tx_pause, prof->tx_ppp,
|
||||
prof->rx_pause, prof->rx_ppp);
|
||||
if (err) {
|
||||
en_err(priv, "Failed setting port general configurations "
|
||||
"for port %d, with error %d\n", priv->port, err);
|
||||
en_err(priv, "Failed setting port general configurations for port %d, with error %d\n",
|
||||
priv->port, err);
|
||||
goto out;
|
||||
}
|
||||
|
||||
|
||||
@@ -270,13 +270,11 @@ static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv)
|
||||
ring->actual_size,
|
||||
GFP_KERNEL)) {
|
||||
if (ring->actual_size < MLX4_EN_MIN_RX_SIZE) {
|
||||
en_err(priv, "Failed to allocate "
|
||||
"enough rx buffers\n");
|
||||
en_err(priv, "Failed to allocate enough rx buffers\n");
|
||||
return -ENOMEM;
|
||||
} else {
|
||||
new_size = rounddown_pow_of_two(ring->actual_size);
|
||||
en_warn(priv, "Only %d buffers allocated "
|
||||
"reducing ring size to %d",
|
||||
en_warn(priv, "Only %d buffers allocated reducing ring size to %d\n",
|
||||
ring->actual_size, new_size);
|
||||
goto reduce_rings;
|
||||
}
|
||||
@@ -685,10 +683,9 @@ int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int bud
|
||||
/* Drop packet on bad receive or bad checksum */
|
||||
if (unlikely((cqe->owner_sr_opcode & MLX4_CQE_OPCODE_MASK) ==
|
||||
MLX4_CQE_OPCODE_ERROR)) {
|
||||
en_err(priv, "CQE completed in error - vendor "
|
||||
"syndrom:%d syndrom:%d\n",
|
||||
((struct mlx4_err_cqe *) cqe)->vendor_err_syndrome,
|
||||
((struct mlx4_err_cqe *) cqe)->syndrome);
|
||||
en_err(priv, "CQE completed in error - vendor syndrom:%d syndrom:%d\n",
|
||||
((struct mlx4_err_cqe *)cqe)->vendor_err_syndrome,
|
||||
((struct mlx4_err_cqe *)cqe)->syndrome);
|
||||
goto next;
|
||||
}
|
||||
if (unlikely(cqe->badfcs_enc & MLX4_CQE_BAD_FCS)) {
|
||||
@@ -944,8 +941,8 @@ void mlx4_en_calc_rx_buf(struct net_device *dev)
|
||||
priv->rx_skb_size = eff_mtu;
|
||||
priv->log_rx_info = ROUNDUP_LOG2(i * sizeof(struct mlx4_en_rx_alloc));
|
||||
|
||||
en_dbg(DRV, priv, "Rx buffer scatter-list (effective-mtu:%d "
|
||||
"num_frags:%d):\n", eff_mtu, priv->num_frags);
|
||||
en_dbg(DRV, priv, "Rx buffer scatter-list (effective-mtu:%d num_frags:%d):\n",
|
||||
eff_mtu, priv->num_frags);
|
||||
for (i = 0; i < priv->num_frags; i++) {
|
||||
en_err(priv,
|
||||
" frag:%d - size:%d prefix:%d align:%d stride:%d\n",
|
||||
|
||||
@@ -108,9 +108,9 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
|
||||
|
||||
ring->buf = ring->wqres.buf.direct.buf;
|
||||
|
||||
en_dbg(DRV, priv, "Allocated TX ring (addr:%p) - buf:%p size:%d "
|
||||
"buf_size:%d dma:%llx\n", ring, ring->buf, ring->size,
|
||||
ring->buf_size, (unsigned long long) ring->wqres.buf.direct.map);
|
||||
en_dbg(DRV, priv, "Allocated TX ring (addr:%p) - buf:%p size:%d buf_size:%d dma:%llx\n",
|
||||
ring, ring->buf, ring->size, ring->buf_size,
|
||||
(unsigned long long) ring->wqres.buf.direct.map);
|
||||
|
||||
ring->qpn = qpn;
|
||||
err = mlx4_qp_alloc(mdev->dev, ring->qpn, &ring->qp);
|
||||
@@ -122,7 +122,7 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
|
||||
|
||||
err = mlx4_bf_alloc(mdev->dev, &ring->bf, node);
|
||||
if (err) {
|
||||
en_dbg(DRV, priv, "working without blueflame (%d)", err);
|
||||
en_dbg(DRV, priv, "working without blueflame (%d)\n", err);
|
||||
ring->bf.uar = &mdev->priv_uar;
|
||||
ring->bf.uar->map = mdev->uar_map;
|
||||
ring->bf_enabled = false;
|
||||
|
||||
@@ -152,14 +152,13 @@ void mlx4_gen_slave_eqe(struct work_struct *work)
|
||||
if (i != dev->caps.function &&
|
||||
master->slave_state[i].active)
|
||||
if (mlx4_GEN_EQE(dev, i, eqe))
|
||||
mlx4_warn(dev, "Failed to "
|
||||
" generate event "
|
||||
"for slave %d\n", i);
|
||||
mlx4_warn(dev, "Failed to generate event for slave %d\n",
|
||||
i);
|
||||
}
|
||||
} else {
|
||||
if (mlx4_GEN_EQE(dev, slave, eqe))
|
||||
mlx4_warn(dev, "Failed to generate event "
|
||||
"for slave %d\n", slave);
|
||||
mlx4_warn(dev, "Failed to generate event for slave %d\n",
|
||||
slave);
|
||||
}
|
||||
++slave_eq->cons;
|
||||
}
|
||||
@@ -177,8 +176,8 @@ static void slave_event(struct mlx4_dev *dev, u8 slave, struct mlx4_eqe *eqe)
|
||||
s_eqe = &slave_eq->event_eqe[slave_eq->prod & (SLAVE_EVENT_EQ_SIZE - 1)];
|
||||
if ((!!(s_eqe->owner & 0x80)) ^
|
||||
(!!(slave_eq->prod & SLAVE_EVENT_EQ_SIZE))) {
|
||||
mlx4_warn(dev, "Master failed to generate an EQE for slave: %d. "
|
||||
"No free EQE on slave events queue\n", slave);
|
||||
mlx4_warn(dev, "Master failed to generate an EQE for slave: %d. No free EQE on slave events queue\n",
|
||||
slave);
|
||||
spin_unlock_irqrestore(&slave_eq->event_lock, flags);
|
||||
return;
|
||||
}
|
||||
@@ -375,9 +374,9 @@ int set_and_calc_slave_port_state(struct mlx4_dev *dev, int slave,
|
||||
}
|
||||
break;
|
||||
default:
|
||||
pr_err("%s: BUG!!! UNKNOWN state: "
|
||||
"slave:%d, port:%d\n", __func__, slave, port);
|
||||
goto out;
|
||||
pr_err("%s: BUG!!! UNKNOWN state: slave:%d, port:%d\n",
|
||||
__func__, slave, port);
|
||||
goto out;
|
||||
}
|
||||
ret = mlx4_get_slave_port_state(dev, slave, port);
|
||||
|
||||
@@ -425,8 +424,8 @@ void mlx4_master_handle_slave_flr(struct work_struct *work)
|
||||
for (i = 0 ; i < dev->num_slaves; i++) {
|
||||
|
||||
if (MLX4_COMM_CMD_FLR == slave_state[i].last_cmd) {
|
||||
mlx4_dbg(dev, "mlx4_handle_slave_flr: "
|
||||
"clean slave: %d\n", i);
|
||||
mlx4_dbg(dev, "mlx4_handle_slave_flr: clean slave: %d\n",
|
||||
i);
|
||||
|
||||
mlx4_delete_all_resources_for_slave(dev, i);
|
||||
/*return the slave to running mode*/
|
||||
@@ -438,8 +437,8 @@ void mlx4_master_handle_slave_flr(struct work_struct *work)
|
||||
err = mlx4_cmd(dev, 0, i, 0, MLX4_CMD_INFORM_FLR_DONE,
|
||||
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
|
||||
if (err)
|
||||
mlx4_warn(dev, "Failed to notify FW on "
|
||||
"FLR done (slave:%d)\n", i);
|
||||
mlx4_warn(dev, "Failed to notify FW on FLR done (slave:%d)\n",
|
||||
i);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -490,9 +489,7 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
be32_to_cpu(eqe->event.qp.qpn)
|
||||
& 0xffffff, &slave);
|
||||
if (ret && ret != -ENOENT) {
|
||||
mlx4_dbg(dev, "QP event %02x(%02x) on "
|
||||
"EQ %d at index %u: could "
|
||||
"not get slave id (%d)\n",
|
||||
mlx4_dbg(dev, "QP event %02x(%02x) on EQ %d at index %u: could not get slave id (%d)\n",
|
||||
eqe->type, eqe->subtype,
|
||||
eq->eqn, eq->cons_index, ret);
|
||||
break;
|
||||
@@ -520,23 +517,19 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
& 0xffffff,
|
||||
&slave);
|
||||
if (ret && ret != -ENOENT) {
|
||||
mlx4_warn(dev, "SRQ event %02x(%02x) "
|
||||
"on EQ %d at index %u: could"
|
||||
" not get slave id (%d)\n",
|
||||
mlx4_warn(dev, "SRQ event %02x(%02x) on EQ %d at index %u: could not get slave id (%d)\n",
|
||||
eqe->type, eqe->subtype,
|
||||
eq->eqn, eq->cons_index, ret);
|
||||
break;
|
||||
}
|
||||
mlx4_warn(dev, "%s: slave:%d, srq_no:0x%x,"
|
||||
" event: %02x(%02x)\n", __func__,
|
||||
slave,
|
||||
mlx4_warn(dev, "%s: slave:%d, srq_no:0x%x, event: %02x(%02x)\n",
|
||||
__func__, slave,
|
||||
be32_to_cpu(eqe->event.srq.srqn),
|
||||
eqe->type, eqe->subtype);
|
||||
|
||||
if (!ret && slave != dev->caps.function) {
|
||||
mlx4_warn(dev, "%s: sending event "
|
||||
"%02x(%02x) to slave:%d\n",
|
||||
__func__, eqe->type,
|
||||
mlx4_warn(dev, "%s: sending event %02x(%02x) to slave:%d\n",
|
||||
__func__, eqe->type,
|
||||
eqe->subtype, slave);
|
||||
mlx4_slave_event(dev, slave, eqe);
|
||||
break;
|
||||
@@ -569,8 +562,7 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
if (dev->caps.port_type[port] == MLX4_PORT_TYPE_ETH) {
|
||||
if (i == mlx4_master_func_num(dev))
|
||||
continue;
|
||||
mlx4_dbg(dev, "%s: Sending MLX4_PORT_CHANGE_SUBTYPE_DOWN"
|
||||
" to slave: %d, port:%d\n",
|
||||
mlx4_dbg(dev, "%s: Sending MLX4_PORT_CHANGE_SUBTYPE_DOWN to slave: %d, port:%d\n",
|
||||
__func__, i, port);
|
||||
s_info = &priv->mfunc.master.vf_oper[slave].vport[port].state;
|
||||
if (IFLA_VF_LINK_STATE_AUTO == s_info->link_state) {
|
||||
@@ -634,11 +626,9 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
be32_to_cpu(eqe->event.cq_err.cqn)
|
||||
& 0xffffff, &slave);
|
||||
if (ret && ret != -ENOENT) {
|
||||
mlx4_dbg(dev, "CQ event %02x(%02x) on "
|
||||
"EQ %d at index %u: could "
|
||||
"not get slave id (%d)\n",
|
||||
eqe->type, eqe->subtype,
|
||||
eq->eqn, eq->cons_index, ret);
|
||||
mlx4_dbg(dev, "CQ event %02x(%02x) on EQ %d at index %u: could not get slave id (%d)\n",
|
||||
eqe->type, eqe->subtype,
|
||||
eq->eqn, eq->cons_index, ret);
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -667,8 +657,7 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
|
||||
case MLX4_EVENT_TYPE_COMM_CHANNEL:
|
||||
if (!mlx4_is_master(dev)) {
|
||||
mlx4_warn(dev, "Received comm channel event "
|
||||
"for non master device\n");
|
||||
mlx4_warn(dev, "Received comm channel event for non master device\n");
|
||||
break;
|
||||
}
|
||||
memcpy(&priv->mfunc.master.comm_arm_bit_vector,
|
||||
@@ -681,8 +670,7 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
case MLX4_EVENT_TYPE_FLR_EVENT:
|
||||
flr_slave = be32_to_cpu(eqe->event.flr_event.slave_id);
|
||||
if (!mlx4_is_master(dev)) {
|
||||
mlx4_warn(dev, "Non-master function received"
|
||||
"FLR event\n");
|
||||
mlx4_warn(dev, "Non-master function received FLR event\n");
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -711,22 +699,17 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
if (eqe->subtype == MLX4_FATAL_WARNING_SUBTYPE_WARMING) {
|
||||
if (mlx4_is_master(dev))
|
||||
for (i = 0; i < dev->num_slaves; i++) {
|
||||
mlx4_dbg(dev, "%s: Sending "
|
||||
"MLX4_FATAL_WARNING_SUBTYPE_WARMING"
|
||||
" to slave: %d\n", __func__, i);
|
||||
mlx4_dbg(dev, "%s: Sending MLX4_FATAL_WARNING_SUBTYPE_WARMING to slave: %d\n",
|
||||
__func__, i);
|
||||
if (i == dev->caps.function)
|
||||
continue;
|
||||
mlx4_slave_event(dev, i, eqe);
|
||||
}
|
||||
mlx4_err(dev, "Temperature Threshold was reached! "
|
||||
"Threshold: %d celsius degrees; "
|
||||
"Current Temperature: %d\n",
|
||||
be16_to_cpu(eqe->event.warming.warning_threshold),
|
||||
be16_to_cpu(eqe->event.warming.current_temperature));
|
||||
mlx4_err(dev, "Temperature Threshold was reached! Threshold: %d celsius degrees; Current Temperature: %d\n",
|
||||
be16_to_cpu(eqe->event.warming.warning_threshold),
|
||||
be16_to_cpu(eqe->event.warming.current_temperature));
|
||||
} else
|
||||
mlx4_warn(dev, "Unhandled event FATAL WARNING (%02x), "
|
||||
"subtype %02x on EQ %d at index %u. owner=%x, "
|
||||
"nent=0x%x, slave=%x, ownership=%s\n",
|
||||
mlx4_warn(dev, "Unhandled event FATAL WARNING (%02x), subtype %02x on EQ %d at index %u. owner=%x, nent=0x%x, slave=%x, ownership=%s\n",
|
||||
eqe->type, eqe->subtype, eq->eqn,
|
||||
eq->cons_index, eqe->owner, eq->nent,
|
||||
eqe->slave_id,
|
||||
@@ -743,9 +726,7 @@ static int mlx4_eq_int(struct mlx4_dev *dev, struct mlx4_eq *eq)
|
||||
case MLX4_EVENT_TYPE_EEC_CATAS_ERROR:
|
||||
case MLX4_EVENT_TYPE_ECC_DETECT:
|
||||
default:
|
||||
mlx4_warn(dev, "Unhandled event %02x(%02x) on EQ %d at "
|
||||
"index %u. owner=%x, nent=0x%x, slave=%x, "
|
||||
"ownership=%s\n",
|
||||
mlx4_warn(dev, "Unhandled event %02x(%02x) on EQ %d at index %u. owner=%x, nent=0x%x, slave=%x, ownership=%s\n",
|
||||
eqe->type, eqe->subtype, eq->eqn,
|
||||
eq->cons_index, eqe->owner, eq->nent,
|
||||
eqe->slave_id,
|
||||
@@ -1088,7 +1069,7 @@ static int mlx4_map_clr_int(struct mlx4_dev *dev)
|
||||
priv->clr_base = ioremap(pci_resource_start(dev->pdev, priv->fw.clr_int_bar) +
|
||||
priv->fw.clr_int_base, MLX4_CLR_INT_SIZE);
|
||||
if (!priv->clr_base) {
|
||||
mlx4_err(dev, "Couldn't map interrupt clear register, aborting.\n");
|
||||
mlx4_err(dev, "Couldn't map interrupt clear register, aborting\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
|
||||
@@ -428,8 +428,7 @@ int mlx4_QUERY_FUNC_CAP(struct mlx4_dev *dev, u32 gen_or_port,
|
||||
} else if (dev->caps.port_type[gen_or_port] == MLX4_PORT_TYPE_IB) {
|
||||
MLX4_GET(field, outbox, QUERY_FUNC_CAP_FLAGS0_OFFSET);
|
||||
if (field & QUERY_FUNC_CAP_FLAGS0_FORCE_PHY_WQE_GID) {
|
||||
mlx4_err(dev, "phy_wqe_gid is "
|
||||
"enforced on this ib port\n");
|
||||
mlx4_err(dev, "phy_wqe_gid is enforced on this ib port\n");
|
||||
err = -EPROTONOSUPPORT;
|
||||
goto out;
|
||||
}
|
||||
@@ -1054,10 +1053,10 @@ int mlx4_map_cmd(struct mlx4_dev *dev, u16 op, struct mlx4_icm *icm, u64 virt)
|
||||
*/
|
||||
lg = ffs(mlx4_icm_addr(&iter) | mlx4_icm_size(&iter)) - 1;
|
||||
if (lg < MLX4_ICM_PAGE_SHIFT) {
|
||||
mlx4_warn(dev, "Got FW area not aligned to %d (%llx/%lx).\n",
|
||||
MLX4_ICM_PAGE_SIZE,
|
||||
(unsigned long long) mlx4_icm_addr(&iter),
|
||||
mlx4_icm_size(&iter));
|
||||
mlx4_warn(dev, "Got FW area not aligned to %d (%llx/%lx)\n",
|
||||
MLX4_ICM_PAGE_SIZE,
|
||||
(unsigned long long) mlx4_icm_addr(&iter),
|
||||
mlx4_icm_size(&iter));
|
||||
err = -EINVAL;
|
||||
goto out;
|
||||
}
|
||||
@@ -1093,14 +1092,14 @@ int mlx4_map_cmd(struct mlx4_dev *dev, u16 op, struct mlx4_icm *icm, u64 virt)
|
||||
|
||||
switch (op) {
|
||||
case MLX4_CMD_MAP_FA:
|
||||
mlx4_dbg(dev, "Mapped %d chunks/%d KB for FW.\n", tc, ts);
|
||||
mlx4_dbg(dev, "Mapped %d chunks/%d KB for FW\n", tc, ts);
|
||||
break;
|
||||
case MLX4_CMD_MAP_ICM_AUX:
|
||||
mlx4_dbg(dev, "Mapped %d chunks/%d KB for ICM aux.\n", tc, ts);
|
||||
mlx4_dbg(dev, "Mapped %d chunks/%d KB for ICM aux\n", tc, ts);
|
||||
break;
|
||||
case MLX4_CMD_MAP_ICM:
|
||||
mlx4_dbg(dev, "Mapped %d chunks/%d KB at %llx for ICM.\n",
|
||||
tc, ts, (unsigned long long) virt - (ts << 10));
|
||||
mlx4_dbg(dev, "Mapped %d chunks/%d KB at %llx for ICM\n",
|
||||
tc, ts, (unsigned long long) virt - (ts << 10));
|
||||
break;
|
||||
}
|
||||
|
||||
@@ -1186,14 +1185,13 @@ int mlx4_QUERY_FW(struct mlx4_dev *dev)
|
||||
MLX4_GET(cmd_if_rev, outbox, QUERY_FW_CMD_IF_REV_OFFSET);
|
||||
if (cmd_if_rev < MLX4_COMMAND_INTERFACE_MIN_REV ||
|
||||
cmd_if_rev > MLX4_COMMAND_INTERFACE_MAX_REV) {
|
||||
mlx4_err(dev, "Installed FW has unsupported "
|
||||
"command interface revision %d.\n",
|
||||
mlx4_err(dev, "Installed FW has unsupported command interface revision %d\n",
|
||||
cmd_if_rev);
|
||||
mlx4_err(dev, "(Installed FW version is %d.%d.%03d)\n",
|
||||
(int) (dev->caps.fw_ver >> 32),
|
||||
(int) (dev->caps.fw_ver >> 16) & 0xffff,
|
||||
(int) dev->caps.fw_ver & 0xffff);
|
||||
mlx4_err(dev, "This driver version supports only revisions %d to %d.\n",
|
||||
mlx4_err(dev, "This driver version supports only revisions %d to %d\n",
|
||||
MLX4_COMMAND_INTERFACE_MIN_REV, MLX4_COMMAND_INTERFACE_MAX_REV);
|
||||
err = -ENODEV;
|
||||
goto out;
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@@ -638,7 +638,7 @@ static int find_entry(struct mlx4_dev *dev, u8 port,
|
||||
|
||||
if (!(be32_to_cpu(mgm->members_count) & 0xffffff)) {
|
||||
if (*index != hash) {
|
||||
mlx4_err(dev, "Found zero MGID in AMGM.\n");
|
||||
mlx4_err(dev, "Found zero MGID in AMGM\n");
|
||||
err = -EINVAL;
|
||||
}
|
||||
return err;
|
||||
@@ -874,7 +874,7 @@ static void mlx4_err_rule(struct mlx4_dev *dev, char *str,
|
||||
mlx4_err(dev, "%s", buf);
|
||||
|
||||
if (len >= BUF_SIZE)
|
||||
mlx4_err(dev, "Network rule error message was truncated, print buffer is too small.\n");
|
||||
mlx4_err(dev, "Network rule error message was truncated, print buffer is too small\n");
|
||||
}
|
||||
|
||||
int mlx4_flow_attach(struct mlx4_dev *dev,
|
||||
@@ -905,10 +905,10 @@ int mlx4_flow_attach(struct mlx4_dev *dev,
|
||||
ret = mlx4_QP_FLOW_STEERING_ATTACH(dev, mailbox, size >> 2, reg_id);
|
||||
if (ret == -ENOMEM)
|
||||
mlx4_err_rule(dev,
|
||||
"mcg table is full. Fail to register network rule.\n",
|
||||
"mcg table is full. Fail to register network rule\n",
|
||||
rule);
|
||||
else if (ret)
|
||||
mlx4_err_rule(dev, "Fail to register network rule.\n", rule);
|
||||
mlx4_err_rule(dev, "Fail to register network rule\n", rule);
|
||||
|
||||
mlx4_free_cmd_mailbox(dev, mailbox);
|
||||
|
||||
@@ -994,7 +994,7 @@ int mlx4_qp_attach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
|
||||
|
||||
members_count = be32_to_cpu(mgm->members_count) & 0xffffff;
|
||||
if (members_count == dev->caps.num_qp_per_mgm) {
|
||||
mlx4_err(dev, "MGM at index %x is full.\n", index);
|
||||
mlx4_err(dev, "MGM at index %x is full\n", index);
|
||||
err = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
@@ -1042,7 +1042,7 @@ out:
|
||||
}
|
||||
if (err && link && index != -1) {
|
||||
if (index < dev->caps.num_mgms)
|
||||
mlx4_warn(dev, "Got AMGM index %d < %d",
|
||||
mlx4_warn(dev, "Got AMGM index %d < %d\n",
|
||||
index, dev->caps.num_mgms);
|
||||
else
|
||||
mlx4_bitmap_free(&priv->mcg_table.bitmap,
|
||||
@@ -1133,7 +1133,7 @@ int mlx4_qp_detach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
|
||||
|
||||
if (amgm_index) {
|
||||
if (amgm_index < dev->caps.num_mgms)
|
||||
mlx4_warn(dev, "MGM entry %d had AMGM index %d < %d",
|
||||
mlx4_warn(dev, "MGM entry %d had AMGM index %d < %d\n",
|
||||
index, amgm_index, dev->caps.num_mgms);
|
||||
else
|
||||
mlx4_bitmap_free(&priv->mcg_table.bitmap,
|
||||
@@ -1153,7 +1153,7 @@ int mlx4_qp_detach_common(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
|
||||
goto out;
|
||||
|
||||
if (index < dev->caps.num_mgms)
|
||||
mlx4_warn(dev, "entry %d had next AMGM index %d < %d",
|
||||
mlx4_warn(dev, "entry %d had next AMGM index %d < %d\n",
|
||||
prev, index, dev->caps.num_mgms);
|
||||
else
|
||||
mlx4_bitmap_free(&priv->mcg_table.bitmap,
|
||||
|
||||
@@ -216,18 +216,19 @@ extern int mlx4_debug_level;
|
||||
#define mlx4_debug_level (0)
|
||||
#endif /* CONFIG_MLX4_DEBUG */
|
||||
|
||||
#define mlx4_dbg(mdev, format, arg...) \
|
||||
#define mlx4_dbg(mdev, format, ...) \
|
||||
do { \
|
||||
if (mlx4_debug_level) \
|
||||
dev_printk(KERN_DEBUG, &mdev->pdev->dev, format, ##arg); \
|
||||
dev_printk(KERN_DEBUG, &(mdev)->pdev->dev, format, \
|
||||
##__VA_ARGS__); \
|
||||
} while (0)
|
||||
|
||||
#define mlx4_err(mdev, format, arg...) \
|
||||
dev_err(&mdev->pdev->dev, format, ##arg)
|
||||
#define mlx4_info(mdev, format, arg...) \
|
||||
dev_info(&mdev->pdev->dev, format, ##arg)
|
||||
#define mlx4_warn(mdev, format, arg...) \
|
||||
dev_warn(&mdev->pdev->dev, format, ##arg)
|
||||
#define mlx4_err(mdev, format, ...) \
|
||||
dev_err(&(mdev)->pdev->dev, format, ##__VA_ARGS__)
|
||||
#define mlx4_info(mdev, format, ...) \
|
||||
dev_info(&(mdev)->pdev->dev, format, ##__VA_ARGS__)
|
||||
#define mlx4_warn(mdev, format, ...) \
|
||||
dev_warn(&(mdev)->pdev->dev, format, ##__VA_ARGS__)
|
||||
|
||||
extern int mlx4_log_num_mgm_entry_size;
|
||||
extern int log_mtts_per_seg;
|
||||
|
||||
@@ -830,26 +830,26 @@ __printf(3, 4)
|
||||
int en_print(const char *level, const struct mlx4_en_priv *priv,
|
||||
const char *format, ...);
|
||||
|
||||
#define en_dbg(mlevel, priv, format, arg...) \
|
||||
do { \
|
||||
if (NETIF_MSG_##mlevel & priv->msg_enable) \
|
||||
en_print(KERN_DEBUG, priv, format, ##arg); \
|
||||
#define en_dbg(mlevel, priv, format, ...) \
|
||||
do { \
|
||||
if (NETIF_MSG_##mlevel & (priv)->msg_enable) \
|
||||
en_print(KERN_DEBUG, priv, format, ##__VA_ARGS__); \
|
||||
} while (0)
|
||||
#define en_warn(priv, format, arg...) \
|
||||
en_print(KERN_WARNING, priv, format, ##arg)
|
||||
#define en_err(priv, format, arg...) \
|
||||
en_print(KERN_ERR, priv, format, ##arg)
|
||||
#define en_info(priv, format, arg...) \
|
||||
en_print(KERN_INFO, priv, format, ## arg)
|
||||
#define en_warn(priv, format, ...) \
|
||||
en_print(KERN_WARNING, priv, format, ##__VA_ARGS__)
|
||||
#define en_err(priv, format, ...) \
|
||||
en_print(KERN_ERR, priv, format, ##__VA_ARGS__)
|
||||
#define en_info(priv, format, ...) \
|
||||
en_print(KERN_INFO, priv, format, ##__VA_ARGS__)
|
||||
|
||||
#define mlx4_err(mdev, format, arg...) \
|
||||
pr_err("%s %s: " format, DRV_NAME, \
|
||||
dev_name(&mdev->pdev->dev), ##arg)
|
||||
#define mlx4_info(mdev, format, arg...) \
|
||||
pr_info("%s %s: " format, DRV_NAME, \
|
||||
dev_name(&mdev->pdev->dev), ##arg)
|
||||
#define mlx4_warn(mdev, format, arg...) \
|
||||
pr_warning("%s %s: " format, DRV_NAME, \
|
||||
dev_name(&mdev->pdev->dev), ##arg)
|
||||
#define mlx4_err(mdev, format, ...) \
|
||||
pr_err(DRV_NAME " %s: " format, \
|
||||
dev_name(&(mdev)->pdev->dev), ##__VA_ARGS__)
|
||||
#define mlx4_info(mdev, format, ...) \
|
||||
pr_info(DRV_NAME " %s: " format, \
|
||||
dev_name(&(mdev)->pdev->dev), ##__VA_ARGS__)
|
||||
#define mlx4_warn(mdev, format, ...) \
|
||||
pr_warn(DRV_NAME " %s: " format, \
|
||||
dev_name(&(mdev)->pdev->dev), ##__VA_ARGS__)
|
||||
|
||||
#endif
|
||||
|
||||
@@ -250,8 +250,8 @@ static void mlx4_free_mtt_range(struct mlx4_dev *dev, u32 offset, int order)
|
||||
MLX4_CMD_TIME_CLASS_A,
|
||||
MLX4_CMD_WRAPPED);
|
||||
if (err)
|
||||
mlx4_warn(dev, "Failed to free mtt range at:"
|
||||
"%d order:%d\n", offset, order);
|
||||
mlx4_warn(dev, "Failed to free mtt range at:%d order:%d\n",
|
||||
offset, order);
|
||||
return;
|
||||
}
|
||||
__mlx4_free_mtt_range(dev, offset, order);
|
||||
@@ -436,8 +436,8 @@ static int mlx4_mr_free_reserved(struct mlx4_dev *dev, struct mlx4_mr *mr)
|
||||
key_to_hw_index(mr->key) &
|
||||
(dev->caps.num_mpts - 1));
|
||||
if (err) {
|
||||
mlx4_warn(dev, "HW2SW_MPT failed (%d),", err);
|
||||
mlx4_warn(dev, "MR has MWs bound to it.\n");
|
||||
mlx4_warn(dev, "HW2SW_MPT failed (%d), MR has MWs bound to it\n",
|
||||
err);
|
||||
return err;
|
||||
}
|
||||
|
||||
@@ -773,7 +773,7 @@ int mlx4_init_mr_table(struct mlx4_dev *dev)
|
||||
mlx4_alloc_mtt_range(dev,
|
||||
fls(dev->caps.reserved_mtts - 1));
|
||||
if (priv->reserved_mtts < 0) {
|
||||
mlx4_warn(dev, "MTT table of order %u is too small.\n",
|
||||
mlx4_warn(dev, "MTT table of order %u is too small\n",
|
||||
mr_table->mtt_buddy.max_order);
|
||||
err = -ENOMEM;
|
||||
goto err_reserve_mtts;
|
||||
@@ -954,8 +954,8 @@ void mlx4_fmr_unmap(struct mlx4_dev *dev, struct mlx4_fmr *fmr,
|
||||
mailbox = mlx4_alloc_cmd_mailbox(dev);
|
||||
if (IS_ERR(mailbox)) {
|
||||
err = PTR_ERR(mailbox);
|
||||
printk(KERN_WARNING "mlx4_ib: mlx4_alloc_cmd_mailbox"
|
||||
" failed (%d)\n", err);
|
||||
printk(KERN_WARNING "mlx4_ib: mlx4_alloc_cmd_mailbox failed (%d)\n",
|
||||
err);
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
@@ -244,8 +244,8 @@ void __mlx4_unregister_mac(struct mlx4_dev *dev, u8 port, u64 mac)
|
||||
if (validate_index(dev, table, index))
|
||||
goto out;
|
||||
if (--table->refs[index]) {
|
||||
mlx4_dbg(dev, "Have more references for index %d,"
|
||||
"no need to modify mac table\n", index);
|
||||
mlx4_dbg(dev, "Have more references for index %d, no need to modify mac table\n",
|
||||
index);
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -443,9 +443,8 @@ void __mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, u16 vlan)
|
||||
}
|
||||
|
||||
if (--table->refs[index]) {
|
||||
mlx4_dbg(dev, "Have %d more references for index %d,"
|
||||
"no need to modify vlan table\n", table->refs[index],
|
||||
index);
|
||||
mlx4_dbg(dev, "Have %d more references for index %d, no need to modify vlan table\n",
|
||||
table->refs[index], index);
|
||||
goto out;
|
||||
}
|
||||
table->entries[index] = 0;
|
||||
@@ -706,8 +705,7 @@ static int mlx4_common_set_port(struct mlx4_dev *dev, int slave, u32 in_mod,
|
||||
if (!memcmp(gid_entry_mbox->raw, gid_entry_tbl->raw,
|
||||
sizeof(gid_entry_tbl->raw))) {
|
||||
/* found duplicate */
|
||||
mlx4_warn(dev, "requested gid entry for slave:%d "
|
||||
"is a duplicate of gid at index %d\n",
|
||||
mlx4_warn(dev, "requested gid entry for slave:%d is a duplicate of gid at index %d\n",
|
||||
slave, i);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@@ -164,18 +164,17 @@ u64 mlx4_make_profile(struct mlx4_dev *dev,
|
||||
}
|
||||
|
||||
if (total_size > dev_cap->max_icm_sz) {
|
||||
mlx4_err(dev, "Profile requires 0x%llx bytes; "
|
||||
"won't fit in 0x%llx bytes of context memory.\n",
|
||||
(unsigned long long) total_size,
|
||||
(unsigned long long) dev_cap->max_icm_sz);
|
||||
mlx4_err(dev, "Profile requires 0x%llx bytes; won't fit in 0x%llx bytes of context memory\n",
|
||||
(unsigned long long) total_size,
|
||||
(unsigned long long) dev_cap->max_icm_sz);
|
||||
kfree(profile);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
if (profile[i].size)
|
||||
mlx4_dbg(dev, " profile[%2d] (%6s): 2^%02d entries @ 0x%10llx, "
|
||||
"size 0x%10llx\n",
|
||||
i, res_name[profile[i].type], profile[i].log_num,
|
||||
mlx4_dbg(dev, " profile[%2d] (%6s): 2^%02d entries @ 0x%10llx, size 0x%10llx\n",
|
||||
i, res_name[profile[i].type],
|
||||
profile[i].log_num,
|
||||
(unsigned long long) profile[i].start,
|
||||
(unsigned long long) profile[i].size);
|
||||
}
|
||||
|
||||
@@ -264,8 +264,8 @@ void mlx4_qp_release_range(struct mlx4_dev *dev, int base_qpn, int cnt)
|
||||
MLX4_CMD_FREE_RES,
|
||||
MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
|
||||
if (err) {
|
||||
mlx4_warn(dev, "Failed to release qp range"
|
||||
" base:%d cnt:%d\n", base_qpn, cnt);
|
||||
mlx4_warn(dev, "Failed to release qp range base:%d cnt:%d\n",
|
||||
base_qpn, cnt);
|
||||
}
|
||||
} else
|
||||
__mlx4_qp_release_range(dev, base_qpn, cnt);
|
||||
@@ -577,8 +577,7 @@ int mlx4_qp_to_ready(struct mlx4_dev *dev, struct mlx4_mtt *mtt,
|
||||
err = mlx4_qp_modify(dev, mtt, states[i], states[i + 1],
|
||||
context, 0, 0, qp);
|
||||
if (err) {
|
||||
mlx4_err(dev, "Failed to bring QP to state: "
|
||||
"%d with error: %d\n",
|
||||
mlx4_err(dev, "Failed to bring QP to state: %d with error: %d\n",
|
||||
states[i + 1], err);
|
||||
return err;
|
||||
}
|
||||
|
||||
@@ -72,8 +72,7 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
hca_header = kmalloc(256, GFP_KERNEL);
|
||||
if (!hca_header) {
|
||||
err = -ENOMEM;
|
||||
mlx4_err(dev, "Couldn't allocate memory to save HCA "
|
||||
"PCI header, aborting.\n");
|
||||
mlx4_err(dev, "Couldn't allocate memory to save HCA PCI header, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -84,8 +83,7 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
continue;
|
||||
if (pci_read_config_dword(dev->pdev, i * 4, hca_header + i)) {
|
||||
err = -ENODEV;
|
||||
mlx4_err(dev, "Couldn't save HCA "
|
||||
"PCI header, aborting.\n");
|
||||
mlx4_err(dev, "Couldn't save HCA PCI header, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
@@ -94,7 +92,7 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
MLX4_RESET_SIZE);
|
||||
if (!reset) {
|
||||
err = -ENOMEM;
|
||||
mlx4_err(dev, "Couldn't map HCA reset register, aborting.\n");
|
||||
mlx4_err(dev, "Couldn't map HCA reset register, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -133,8 +131,7 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
|
||||
if (vendor == 0xffff) {
|
||||
err = -ENODEV;
|
||||
mlx4_err(dev, "PCI device did not come back after reset, "
|
||||
"aborting.\n");
|
||||
mlx4_err(dev, "PCI device did not come back after reset, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -144,16 +141,14 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
if (pcie_capability_write_word(dev->pdev, PCI_EXP_DEVCTL,
|
||||
devctl)) {
|
||||
err = -ENODEV;
|
||||
mlx4_err(dev, "Couldn't restore HCA PCI Express "
|
||||
"Device Control register, aborting.\n");
|
||||
mlx4_err(dev, "Couldn't restore HCA PCI Express Device Control register, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
linkctl = hca_header[(pcie_cap + PCI_EXP_LNKCTL) / 4];
|
||||
if (pcie_capability_write_word(dev->pdev, PCI_EXP_LNKCTL,
|
||||
linkctl)) {
|
||||
err = -ENODEV;
|
||||
mlx4_err(dev, "Couldn't restore HCA PCI Express "
|
||||
"Link control register, aborting.\n");
|
||||
mlx4_err(dev, "Couldn't restore HCA PCI Express Link control register, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
@@ -164,8 +159,8 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
|
||||
if (pci_write_config_dword(dev->pdev, i * 4, hca_header[i])) {
|
||||
err = -ENODEV;
|
||||
mlx4_err(dev, "Couldn't restore HCA reg %x, "
|
||||
"aborting.\n", i);
|
||||
mlx4_err(dev, "Couldn't restore HCA reg %x, aborting\n",
|
||||
i);
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
@@ -173,8 +168,7 @@ int mlx4_reset(struct mlx4_dev *dev)
|
||||
if (pci_write_config_dword(dev->pdev, PCI_COMMAND,
|
||||
hca_header[PCI_COMMAND / 4])) {
|
||||
err = -ENODEV;
|
||||
mlx4_err(dev, "Couldn't restore HCA COMMAND, "
|
||||
"aborting.\n");
|
||||
mlx4_err(dev, "Couldn't restore HCA COMMAND, aborting\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
|
||||
@@ -3857,7 +3857,7 @@ static int add_eth_header(struct mlx4_dev *dev, int slave,
|
||||
}
|
||||
}
|
||||
if (!be_mac) {
|
||||
pr_err("Failed adding eth header to FS rule, Can't find matching MAC for port %d .\n",
|
||||
pr_err("Failed adding eth header to FS rule, Can't find matching MAC for port %d\n",
|
||||
port);
|
||||
return -EINVAL;
|
||||
}
|
||||
@@ -3900,7 +3900,7 @@ int mlx4_QP_FLOW_STEERING_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||
qpn = be32_to_cpu(ctrl->qpn) & 0xffffff;
|
||||
err = get_res(dev, slave, qpn, RES_QP, &rqp);
|
||||
if (err) {
|
||||
pr_err("Steering rule with qpn 0x%x rejected.\n", qpn);
|
||||
pr_err("Steering rule with qpn 0x%x rejected\n", qpn);
|
||||
return err;
|
||||
}
|
||||
rule_header = (struct _rule_hw *)(ctrl + 1);
|
||||
@@ -3918,7 +3918,7 @@ int mlx4_QP_FLOW_STEERING_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||
case MLX4_NET_TRANS_RULE_ID_IPV4:
|
||||
case MLX4_NET_TRANS_RULE_ID_TCP:
|
||||
case MLX4_NET_TRANS_RULE_ID_UDP:
|
||||
pr_warn("Can't attach FS rule without L2 headers, adding L2 header.\n");
|
||||
pr_warn("Can't attach FS rule without L2 headers, adding L2 header\n");
|
||||
if (add_eth_header(dev, slave, inbox, rlist, header_id)) {
|
||||
err = -EINVAL;
|
||||
goto err_put;
|
||||
@@ -3927,7 +3927,7 @@ int mlx4_QP_FLOW_STEERING_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||
sizeof(struct mlx4_net_trans_rule_hw_eth) >> 2;
|
||||
break;
|
||||
default:
|
||||
pr_err("Corrupted mailbox.\n");
|
||||
pr_err("Corrupted mailbox\n");
|
||||
err = -EINVAL;
|
||||
goto err_put;
|
||||
}
|
||||
@@ -3941,7 +3941,7 @@ int mlx4_QP_FLOW_STEERING_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||
|
||||
err = add_res_range(dev, slave, vhcr->out_param, 1, RES_FS_RULE, qpn);
|
||||
if (err) {
|
||||
mlx4_err(dev, "Fail to add flow steering resources.\n ");
|
||||
mlx4_err(dev, "Fail to add flow steering resources\n");
|
||||
/* detach rule*/
|
||||
mlx4_cmd(dev, vhcr->out_param, 0, 0,
|
||||
MLX4_QP_FLOW_STEERING_DETACH, MLX4_CMD_TIME_CLASS_A,
|
||||
@@ -3979,7 +3979,7 @@ int mlx4_QP_FLOW_STEERING_DETACH_wrapper(struct mlx4_dev *dev, int slave,
|
||||
|
||||
err = rem_res_range(dev, slave, vhcr->in_param, 1, RES_FS_RULE, 0);
|
||||
if (err) {
|
||||
mlx4_err(dev, "Fail to remove flow steering resources.\n ");
|
||||
mlx4_err(dev, "Fail to remove flow steering resources\n");
|
||||
goto out;
|
||||
}
|
||||
|
||||
@@ -4108,8 +4108,8 @@ static void rem_slave_qps(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_QP);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_qps: Could not move all qps to busy"
|
||||
"for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_qps: Could not move all qps to busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(qp, tmp, qp_list, com.list) {
|
||||
@@ -4147,10 +4147,8 @@ static void rem_slave_qps(struct mlx4_dev *dev, int slave)
|
||||
MLX4_CMD_TIME_CLASS_A,
|
||||
MLX4_CMD_NATIVE);
|
||||
if (err)
|
||||
mlx4_dbg(dev, "rem_slave_qps: failed"
|
||||
" to move slave %d qpn %d to"
|
||||
" reset\n", slave,
|
||||
qp->local_qpn);
|
||||
mlx4_dbg(dev, "rem_slave_qps: failed to move slave %d qpn %d to reset\n",
|
||||
slave, qp->local_qpn);
|
||||
atomic_dec(&qp->rcq->ref_count);
|
||||
atomic_dec(&qp->scq->ref_count);
|
||||
atomic_dec(&qp->mtt->ref_count);
|
||||
@@ -4184,8 +4182,8 @@ static void rem_slave_srqs(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_SRQ);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_srqs: Could not move all srqs to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_srqs: Could not move all srqs - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(srq, tmp, srq_list, com.list) {
|
||||
@@ -4215,9 +4213,7 @@ static void rem_slave_srqs(struct mlx4_dev *dev, int slave)
|
||||
MLX4_CMD_TIME_CLASS_A,
|
||||
MLX4_CMD_NATIVE);
|
||||
if (err)
|
||||
mlx4_dbg(dev, "rem_slave_srqs: failed"
|
||||
" to move slave %d srq %d to"
|
||||
" SW ownership\n",
|
||||
mlx4_dbg(dev, "rem_slave_srqs: failed to move slave %d srq %d to SW ownership\n",
|
||||
slave, srqn);
|
||||
|
||||
atomic_dec(&srq->mtt->ref_count);
|
||||
@@ -4252,8 +4248,8 @@ static void rem_slave_cqs(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_CQ);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_cqs: Could not move all cqs to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_cqs: Could not move all cqs - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(cq, tmp, cq_list, com.list) {
|
||||
@@ -4283,9 +4279,7 @@ static void rem_slave_cqs(struct mlx4_dev *dev, int slave)
|
||||
MLX4_CMD_TIME_CLASS_A,
|
||||
MLX4_CMD_NATIVE);
|
||||
if (err)
|
||||
mlx4_dbg(dev, "rem_slave_cqs: failed"
|
||||
" to move slave %d cq %d to"
|
||||
" SW ownership\n",
|
||||
mlx4_dbg(dev, "rem_slave_cqs: failed to move slave %d cq %d to SW ownership\n",
|
||||
slave, cqn);
|
||||
atomic_dec(&cq->mtt->ref_count);
|
||||
state = RES_CQ_ALLOCATED;
|
||||
@@ -4317,8 +4311,8 @@ static void rem_slave_mrs(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_MPT);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_mrs: Could not move all mpts to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_mrs: Could not move all mpts - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(mpt, tmp, mpt_list, com.list) {
|
||||
@@ -4353,9 +4347,7 @@ static void rem_slave_mrs(struct mlx4_dev *dev, int slave)
|
||||
MLX4_CMD_TIME_CLASS_A,
|
||||
MLX4_CMD_NATIVE);
|
||||
if (err)
|
||||
mlx4_dbg(dev, "rem_slave_mrs: failed"
|
||||
" to move slave %d mpt %d to"
|
||||
" SW ownership\n",
|
||||
mlx4_dbg(dev, "rem_slave_mrs: failed to move slave %d mpt %d to SW ownership\n",
|
||||
slave, mptn);
|
||||
if (mpt->mtt)
|
||||
atomic_dec(&mpt->mtt->ref_count);
|
||||
@@ -4387,8 +4379,8 @@ static void rem_slave_mtts(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_MTT);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_mtts: Could not move all mtts to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_mtts: Could not move all mtts - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(mtt, tmp, mtt_list, com.list) {
|
||||
@@ -4490,8 +4482,8 @@ static void rem_slave_eqs(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_EQ);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_eqs: Could not move all eqs to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_eqs: Could not move all eqs - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(eq, tmp, eq_list, com.list) {
|
||||
@@ -4523,9 +4515,8 @@ static void rem_slave_eqs(struct mlx4_dev *dev, int slave)
|
||||
MLX4_CMD_TIME_CLASS_A,
|
||||
MLX4_CMD_NATIVE);
|
||||
if (err)
|
||||
mlx4_dbg(dev, "rem_slave_eqs: failed"
|
||||
" to move slave %d eqs %d to"
|
||||
" SW ownership\n", slave, eqn);
|
||||
mlx4_dbg(dev, "rem_slave_eqs: failed to move slave %d eqs %d to SW ownership\n",
|
||||
slave, eqn);
|
||||
mlx4_free_cmd_mailbox(dev, mailbox);
|
||||
atomic_dec(&eq->mtt->ref_count);
|
||||
state = RES_EQ_RESERVED;
|
||||
@@ -4554,8 +4545,8 @@ static void rem_slave_counters(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_COUNTER);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_counters: Could not move all counters to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_counters: Could not move all counters - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(counter, tmp, counter_list, com.list) {
|
||||
@@ -4585,8 +4576,8 @@ static void rem_slave_xrcdns(struct mlx4_dev *dev, int slave)
|
||||
|
||||
err = move_all_busy(dev, slave, RES_XRCD);
|
||||
if (err)
|
||||
mlx4_warn(dev, "rem_slave_xrcdns: Could not move all xrcdns to "
|
||||
"busy for slave %d\n", slave);
|
||||
mlx4_warn(dev, "rem_slave_xrcdns: Could not move all xrcdns - too busy for slave %d\n",
|
||||
slave);
|
||||
|
||||
spin_lock_irq(mlx4_tlock(dev));
|
||||
list_for_each_entry_safe(xrcd, tmp, xrcdn_list, com.list) {
|
||||
@@ -4731,10 +4722,8 @@ void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work)
|
||||
0, MLX4_CMD_UPDATE_QP,
|
||||
MLX4_CMD_TIME_CLASS_C, MLX4_CMD_NATIVE);
|
||||
if (err) {
|
||||
mlx4_info(dev, "UPDATE_QP failed for slave %d, "
|
||||
"port %d, qpn %d (%d)\n",
|
||||
work->slave, port, qp->local_qpn,
|
||||
err);
|
||||
mlx4_info(dev, "UPDATE_QP failed for slave %d, port %d, qpn %d (%d)\n",
|
||||
work->slave, port, qp->local_qpn, err);
|
||||
errors++;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -620,8 +620,8 @@ static int wait_func(struct mlx5_core_dev *dev, struct mlx5_cmd_work_ent *ent)
|
||||
mlx5_command_str(msg_to_opcode(ent->in)),
|
||||
msg_to_opcode(ent->in));
|
||||
}
|
||||
mlx5_core_dbg(dev, "err %d, delivery status %s(%d)\n", err,
|
||||
deliv_status_to_str(ent->status), ent->status);
|
||||
mlx5_core_dbg(dev, "err %d, delivery status %s(%d)\n",
|
||||
err, deliv_status_to_str(ent->status), ent->status);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user