On 8/16/17, 5:05 AM, "[email protected] on behalf of Colin
King" <[email protected] on behalf of
[email protected]> wrote:

>From: Colin Ian King <[email protected]>
>
>Trivial fix to spelling mistakes in the mlx4 driver
>
>Signed-off-by: Colin Ian King <[email protected]>
>---
> drivers/net/ethernet/mellanox/mlx4/cmd.c    | 16 ++++++++--------
> drivers/net/ethernet/mellanox/mlx4/fw_qos.c |  6 +++---
> drivers/net/ethernet/mellanox/mlx4/fw_qos.h | 10 +++++-----
> 3 files changed, 16 insertions(+), 16 deletions(-)
>
>diff --git a/drivers/net/ethernet/mellanox/mlx4/cmd.c
>b/drivers/net/ethernet/mellanox/mlx4/cmd.c
>index 674773b28b2e..6309389b09a7 100644
>--- a/drivers/net/ethernet/mellanox/mlx4/cmd.c
>+++ b/drivers/net/ethernet/mellanox/mlx4/cmd.c
>@@ -1958,19 +1958,19 @@ static void mlx4_allocate_port_vpps(struct
>mlx4_dev *dev, int port)
>       int i;
>       int err;
>       int num_vfs;
>-      u16 availible_vpp;
>+      u16 available_vpp;
>       u8 vpp_param[MLX4_NUM_UP];
>       struct mlx4_qos_manager *port_qos;
>       struct mlx4_priv *priv = mlx4_priv(dev);
> 
>-      err = mlx4_ALLOCATE_VPP_get(dev, port, &availible_vpp, vpp_param);
>+      err = mlx4_ALLOCATE_VPP_get(dev, port, &available_vpp, vpp_param);
>       if (err) {
>-              mlx4_info(dev, "Failed query availible VPPs\n");
>+              mlx4_info(dev, "Failed query available VPPs\n");
>               return;
>       }
> 
>       port_qos = &priv->mfunc.master.qos_ctl[port];
>-      num_vfs = (availible_vpp /
>+      num_vfs = (available_vpp /
>                  bitmap_weight(port_qos->priority_bm, MLX4_NUM_UP));
> 
>       for (i = 0; i < MLX4_NUM_UP; i++) {
>@@ -1985,14 +1985,14 @@ static void mlx4_allocate_port_vpps(struct
>mlx4_dev *dev, int port)
>       }
> 
>       /* Query actual allocated VPP, just to make sure */
>-      err = mlx4_ALLOCATE_VPP_get(dev, port, &availible_vpp, vpp_param);
>+      err = mlx4_ALLOCATE_VPP_get(dev, port, &available_vpp, vpp_param);
>       if (err) {
>-              mlx4_info(dev, "Failed query availible VPPs\n");
>+              mlx4_info(dev, "Failed query available VPPs\n");
>               return;
>       }
> 
>       port_qos->num_of_qos_vfs = num_vfs;
>-      mlx4_dbg(dev, "Port %d Availible VPPs %d\n", port, availible_vpp);
>+      mlx4_dbg(dev, "Port %d Availible VPPs %d\n", port, available_vpp);

One more here, in the text.          ^^^

> 
>       for (i = 0; i < MLX4_NUM_UP; i++)
>               mlx4_dbg(dev, "Port %d UP %d Allocated %d VPPs\n", port, i,
>@@ -2891,7 +2891,7 @@ static int mlx4_set_vport_qos(struct mlx4_priv
>*priv, int slave, int port,
>       memset(vpp_qos, 0, sizeof(struct mlx4_vport_qos_param) * MLX4_NUM_UP);
> 
>       if (slave > port_qos->num_of_qos_vfs) {
>-              mlx4_info(dev, "No availible VPP resources for this VF\n");
>+              mlx4_info(dev, "No available VPP resources for this VF\n");
>               return -EINVAL;
>       }
> 
>diff --git a/drivers/net/ethernet/mellanox/mlx4/fw_qos.c
>b/drivers/net/ethernet/mellanox/mlx4/fw_qos.c
>index 8f2fde0487c4..3a09d7122d3b 100644
>--- a/drivers/net/ethernet/mellanox/mlx4/fw_qos.c
>+++ b/drivers/net/ethernet/mellanox/mlx4/fw_qos.c
>@@ -65,7 +65,7 @@ struct mlx4_set_port_scheduler_context {
> 
> /* Granular Qos (per VF) section */
> struct mlx4_alloc_vpp_param {
>-      __be32 availible_vpp;
>+      __be32 available_vpp;
>       __be32 vpp_p_up[MLX4_NUM_UP];
> };
> 
>@@ -157,7 +157,7 @@ int mlx4_SET_PORT_SCHEDULER(struct mlx4_dev *dev, u8
>port, u8 *tc_tx_bw,
> EXPORT_SYMBOL(mlx4_SET_PORT_SCHEDULER);
> 
> int mlx4_ALLOCATE_VPP_get(struct mlx4_dev *dev, u8 port,
>-                        u16 *availible_vpp, u8 *vpp_p_up)
>+                        u16 *available_vpp, u8 *vpp_p_up)
> {
>       int i;
>       int err;
>@@ -179,7 +179,7 @@ int mlx4_ALLOCATE_VPP_get(struct mlx4_dev *dev, u8
>port,
>               goto out;
> 
>       /* Total number of supported VPPs */
>-      *availible_vpp = (u16)be32_to_cpu(out_param->availible_vpp);
>+      *available_vpp = (u16)be32_to_cpu(out_param->available_vpp);
> 
>       for (i = 0; i < MLX4_NUM_UP; i++)
>               vpp_p_up[i] = (u8)be32_to_cpu(out_param->vpp_p_up[i]);
>diff --git a/drivers/net/ethernet/mellanox/mlx4/fw_qos.h
>b/drivers/net/ethernet/mellanox/mlx4/fw_qos.h
>index ac1f331878e6..582997577a04 100644
>--- a/drivers/net/ethernet/mellanox/mlx4/fw_qos.h
>+++ b/drivers/net/ethernet/mellanox/mlx4/fw_qos.h
>@@ -84,23 +84,23 @@ int mlx4_SET_PORT_PRIO2TC(struct mlx4_dev *dev, u8
>port, u8 *prio2tc);
> int mlx4_SET_PORT_SCHEDULER(struct mlx4_dev *dev, u8 port, u8 *tc_tx_bw,
>                           u8 *pg, u16 *ratelimit);
> /**
>- * mlx4_ALLOCATE_VPP_get - Query port VPP availible resources and
>allocation.
>- * Before distribution of VPPs to priorities, only availible_vpp is
>returned.
>+ * mlx4_ALLOCATE_VPP_get - Query port VPP available resources and
>allocation.
>+ * Before distribution of VPPs to priorities, only available_vpp is
>returned.
>  * After initialization it returns the distribution of VPPs among
>priorities.
>  *
>  * @dev: mlx4_dev.
>  * @port: Physical port number.
>- * @availible_vpp: Pointer to variable where number of availible VPPs is
>stored
>+ * @available_vpp: Pointer to variable where number of available VPPs is
>stored
>  * @vpp_p_up: Distribution of VPPs to priorities is stored in this array
>  *
>  * Returns 0 on success or a negative mlx4_core errno code.
>  **/
> int mlx4_ALLOCATE_VPP_get(struct mlx4_dev *dev, u8 port,
>-                        u16 *availible_vpp, u8 *vpp_p_up);
>+                        u16 *available_vpp, u8 *vpp_p_up);
> /**
>  * mlx4_ALLOCATE_VPP_set - Distribution of VPPs among differnt
>priorities.
>  * The total number of VPPs assigned to all for a port must not exceed
>- * the value reported by availible_vpp in mlx4_ALLOCATE_VPP_get.
>+ * the value reported by available_vpp in mlx4_ALLOCATE_VPP_get.
>  * VPP allocation is allowed only after the port type has been set,
>  * and while no QPs are open for this port.
>  *
>-- 
>2.11.0
>
>--
>To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
>the body of a message to [email protected]
>More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to