net/mlx5: Introduce TSAR manipulation firmware commands
authorMohamad Haj Yahia <mohamad@mellanox.com>
Thu, 11 Aug 2016 08:21:39 +0000 (11:21 +0300)
committerLeon Romanovsky <leon@kernel.org>
Sun, 30 Oct 2016 13:43:12 +0000 (15:43 +0200)
TSAR (stands for Transmit Scheduling ARbiter) is a hardware component
that is responsible for selecting the next entity to serve on the
transmit path.
The arbitration defines the QoS policy between the agents connected to
the TSAR.
The TSAR is a consist two main features:
1) BW Allocation between agents:
The TSAR implements a defecit weighted round robin between the agents.
Each agent attached to the TSAR is assigned with a weight and it is
awarded transmission tokens according to this weight.
2) Rate limer per agent:
Each agent attached to the TSAR is (optionally) assigned with a rate
limit.
TSAR will not allow scheduling for an agent exceeding its defined rate
limit.

In this patch we implement the API of manipulating the TSAR.

Signed-off-by: Mohamad Haj Yahia <mohamad@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
Signed-off-by: Leon Romanovsky <leon@kernel.org>
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
drivers/net/ethernet/mellanox/mlx5/core/mlx5_core.h
drivers/net/ethernet/mellanox/mlx5/core/rl.c
include/linux/mlx5/mlx5_ifc.h

index 1e639f88602165fcb4f4454cf6a54148372e3235..8561102f2563dd2e0b92dafb6d2485c815eff753 100644 (file)
@@ -318,6 +318,8 @@ static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op,
        case MLX5_CMD_OP_SET_FLOW_TABLE_ENTRY:
        case MLX5_CMD_OP_SET_FLOW_TABLE_ROOT:
        case MLX5_CMD_OP_DEALLOC_ENCAP_HEADER:
+       case MLX5_CMD_OP_DESTROY_SCHEDULING_ELEMENT:
+       case MLX5_CMD_OP_DESTROY_QOS_PARA_VPORT:
                return MLX5_CMD_STAT_OK;
 
        case MLX5_CMD_OP_QUERY_HCA_CAP:
@@ -419,11 +421,14 @@ static int mlx5_internal_err_ret_value(struct mlx5_core_dev *dev, u16 op,
        case MLX5_CMD_OP_QUERY_FLOW_TABLE:
        case MLX5_CMD_OP_CREATE_FLOW_GROUP:
        case MLX5_CMD_OP_QUERY_FLOW_GROUP:
-
        case MLX5_CMD_OP_QUERY_FLOW_TABLE_ENTRY:
        case MLX5_CMD_OP_ALLOC_FLOW_COUNTER:
        case MLX5_CMD_OP_QUERY_FLOW_COUNTER:
        case MLX5_CMD_OP_ALLOC_ENCAP_HEADER:
+       case MLX5_CMD_OP_CREATE_SCHEDULING_ELEMENT:
+       case MLX5_CMD_OP_QUERY_SCHEDULING_ELEMENT:
+       case MLX5_CMD_OP_MODIFY_SCHEDULING_ELEMENT:
+       case MLX5_CMD_OP_CREATE_QOS_PARA_VPORT:
                *status = MLX5_DRIVER_STATUS_ABORTED;
                *synd = MLX5_DRIVER_SYND;
                return -EIO;
@@ -580,6 +585,12 @@ const char *mlx5_command_str(int command)
        MLX5_COMMAND_STR_CASE(MODIFY_FLOW_TABLE);
        MLX5_COMMAND_STR_CASE(ALLOC_ENCAP_HEADER);
        MLX5_COMMAND_STR_CASE(DEALLOC_ENCAP_HEADER);
+       MLX5_COMMAND_STR_CASE(CREATE_SCHEDULING_ELEMENT);
+       MLX5_COMMAND_STR_CASE(DESTROY_SCHEDULING_ELEMENT);
+       MLX5_COMMAND_STR_CASE(QUERY_SCHEDULING_ELEMENT);
+       MLX5_COMMAND_STR_CASE(MODIFY_SCHEDULING_ELEMENT);
+       MLX5_COMMAND_STR_CASE(CREATE_QOS_PARA_VPORT);
+       MLX5_COMMAND_STR_CASE(DESTROY_QOS_PARA_VPORT);
        default: return "unknown command opcode";
        }
 }
index 3d0cfb9f18f99154e6f904625a41ff457c4e2508..bf431715172ca8ab68c5b03d23c2c6617ee2777e 100644 (file)
@@ -91,6 +91,13 @@ int mlx5_core_sriov_configure(struct pci_dev *dev, int num_vfs);
 bool mlx5_sriov_is_enabled(struct mlx5_core_dev *dev);
 int mlx5_core_enable_hca(struct mlx5_core_dev *dev, u16 func_id);
 int mlx5_core_disable_hca(struct mlx5_core_dev *dev, u16 func_id);
+int mlx5_create_scheduling_element_cmd(struct mlx5_core_dev *dev, u8 hierarchy,
+                                      void *context, u32 *element_id);
+int mlx5_modify_scheduling_element_cmd(struct mlx5_core_dev *dev, u8 hierarchy,
+                                      void *context, u32 element_id,
+                                      u32 modify_bitmask);
+int mlx5_destroy_scheduling_element_cmd(struct mlx5_core_dev *dev, u8 hierarchy,
+                                       u32 element_id);
 int mlx5_wait_for_vf_pages(struct mlx5_core_dev *dev);
 cycle_t mlx5_read_internal_timer(struct mlx5_core_dev *dev);
 u32 mlx5_get_msix_vec(struct mlx5_core_dev *dev, int vecidx);
index 104902a93a0b577f22ecd43824c6c42b5a9b059d..e651e4c02867740d35c07bfcf485860f26ad6409 100644 (file)
 #include <linux/mlx5/cmd.h>
 #include "mlx5_core.h"
 
+/* Scheduling element fw management */
+int mlx5_create_scheduling_element_cmd(struct mlx5_core_dev *dev, u8 hierarchy,
+                                      void *ctx, u32 *element_id)
+{
+       u32 in[MLX5_ST_SZ_DW(create_scheduling_element_in)]  = {0};
+       u32 out[MLX5_ST_SZ_DW(create_scheduling_element_in)] = {0};
+       void *schedc;
+       int err;
+
+       schedc = MLX5_ADDR_OF(create_scheduling_element_in, in,
+                             scheduling_context);
+       MLX5_SET(create_scheduling_element_in, in, opcode,
+                MLX5_CMD_OP_CREATE_SCHEDULING_ELEMENT);
+       MLX5_SET(create_scheduling_element_in, in, scheduling_hierarchy,
+                hierarchy);
+       memcpy(schedc, ctx, MLX5_ST_SZ_BYTES(scheduling_context));
+
+       err = mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out));
+       if (err)
+               return err;
+
+       *element_id = MLX5_GET(create_scheduling_element_out, out,
+                              scheduling_element_id);
+       return 0;
+}
+
+int mlx5_modify_scheduling_element_cmd(struct mlx5_core_dev *dev, u8 hierarchy,
+                                      void *ctx, u32 element_id,
+                                      u32 modify_bitmask)
+{
+       u32 in[MLX5_ST_SZ_DW(modify_scheduling_element_in)]  = {0};
+       u32 out[MLX5_ST_SZ_DW(modify_scheduling_element_in)] = {0};
+       void *schedc;
+
+       schedc = MLX5_ADDR_OF(modify_scheduling_element_in, in,
+                             scheduling_context);
+       MLX5_SET(modify_scheduling_element_in, in, opcode,
+                MLX5_CMD_OP_MODIFY_SCHEDULING_ELEMENT);
+       MLX5_SET(modify_scheduling_element_in, in, scheduling_element_id,
+                element_id);
+       MLX5_SET(modify_scheduling_element_in, in, modify_bitmask,
+                modify_bitmask);
+       MLX5_SET(modify_scheduling_element_in, in, scheduling_hierarchy,
+                hierarchy);
+       memcpy(schedc, ctx, MLX5_ST_SZ_BYTES(scheduling_context));
+
+       return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out));
+}
+
+int mlx5_destroy_scheduling_element_cmd(struct mlx5_core_dev *dev, u8 hierarchy,
+                                       u32 element_id)
+{
+       u32 in[MLX5_ST_SZ_DW(destroy_scheduling_element_in)]  = {0};
+       u32 out[MLX5_ST_SZ_DW(destroy_scheduling_element_in)] = {0};
+
+       MLX5_SET(destroy_scheduling_element_in, in, opcode,
+                MLX5_CMD_OP_DESTROY_SCHEDULING_ELEMENT);
+       MLX5_SET(destroy_scheduling_element_in, in, scheduling_element_id,
+                element_id);
+       MLX5_SET(destroy_scheduling_element_in, in, scheduling_hierarchy,
+                hierarchy);
+
+       return mlx5_cmd_exec(dev, in, sizeof(in), out, sizeof(out));
+}
+
 /* Finds an entry where we can register the given rate
  * If the rate already exists, return the entry where it is registered,
  * otherwise return the first available entry.
index 12f72e45a3f0be0ead36aff09edbac70b66a746c..2632cb2caf1005c6886471dbd1c04bb62a1c98cf 100644 (file)
@@ -145,6 +145,12 @@ enum {
        MLX5_CMD_OP_QUERY_Q_COUNTER               = 0x773,
        MLX5_CMD_OP_SET_RATE_LIMIT                = 0x780,
        MLX5_CMD_OP_QUERY_RATE_LIMIT              = 0x781,
+       MLX5_CMD_OP_CREATE_SCHEDULING_ELEMENT      = 0x782,
+       MLX5_CMD_OP_DESTROY_SCHEDULING_ELEMENT     = 0x783,
+       MLX5_CMD_OP_QUERY_SCHEDULING_ELEMENT       = 0x784,
+       MLX5_CMD_OP_MODIFY_SCHEDULING_ELEMENT      = 0x785,
+       MLX5_CMD_OP_CREATE_QOS_PARA_VPORT         = 0x786,
+       MLX5_CMD_OP_DESTROY_QOS_PARA_VPORT        = 0x787,
        MLX5_CMD_OP_ALLOC_PD                      = 0x800,
        MLX5_CMD_OP_DEALLOC_PD                    = 0x801,
        MLX5_CMD_OP_ALLOC_UAR                     = 0x802,
@@ -537,13 +543,27 @@ struct mlx5_ifc_e_switch_cap_bits {
 
 struct mlx5_ifc_qos_cap_bits {
        u8         packet_pacing[0x1];
-       u8         reserved_0[0x1f];
-       u8         reserved_1[0x20];
+       u8         esw_scheduling[0x1];
+       u8         reserved_at_2[0x1e];
+
+       u8         reserved_at_20[0x20];
+
        u8         packet_pacing_max_rate[0x20];
+
        u8         packet_pacing_min_rate[0x20];
-       u8         reserved_2[0x10];
+
+       u8         reserved_at_80[0x10];
        u8         packet_pacing_rate_table_size[0x10];
-       u8         reserved_3[0x760];
+
+       u8         esw_element_type[0x10];
+       u8         esw_tsar_type[0x10];
+
+       u8         reserved_at_c0[0x10];
+       u8         max_qos_para_vport[0x10];
+
+       u8         max_tsar_bw_share[0x20];
+
+       u8         reserved_at_100[0x700];
 };
 
 struct mlx5_ifc_per_protocol_networking_offload_caps_bits {
@@ -2333,6 +2353,30 @@ struct mlx5_ifc_sqc_bits {
        struct mlx5_ifc_wq_bits wq;
 };
 
+enum {
+       SCHEDULING_CONTEXT_ELEMENT_TYPE_TSAR = 0x0,
+       SCHEDULING_CONTEXT_ELEMENT_TYPE_VPORT = 0x1,
+       SCHEDULING_CONTEXT_ELEMENT_TYPE_VPORT_TC = 0x2,
+       SCHEDULING_CONTEXT_ELEMENT_TYPE_PARA_VPORT_TC = 0x3,
+};
+
+struct mlx5_ifc_scheduling_context_bits {
+       u8         element_type[0x8];
+       u8         reserved_at_8[0x18];
+
+       u8         element_attributes[0x20];
+
+       u8         parent_element_id[0x20];
+
+       u8         reserved_at_60[0x40];
+
+       u8         bw_share[0x20];
+
+       u8         max_average_bw[0x20];
+
+       u8         reserved_at_e0[0x120];
+};
+
 struct mlx5_ifc_rqtc_bits {
        u8         reserved_at_0[0xa0];
 
@@ -2920,6 +2964,29 @@ struct mlx5_ifc_register_loopback_control_bits {
        u8         reserved_at_20[0x60];
 };
 
+struct mlx5_ifc_vport_tc_element_bits {
+       u8         traffic_class[0x4];
+       u8         reserved_at_4[0xc];
+       u8         vport_number[0x10];
+};
+
+struct mlx5_ifc_vport_element_bits {
+       u8         reserved_at_0[0x10];
+       u8         vport_number[0x10];
+};
+
+enum {
+       TSAR_ELEMENT_TSAR_TYPE_DWRR = 0x0,
+       TSAR_ELEMENT_TSAR_TYPE_ROUND_ROBIN = 0x1,
+       TSAR_ELEMENT_TSAR_TYPE_ETS = 0x2,
+};
+
+struct mlx5_ifc_tsar_element_bits {
+       u8         reserved_at_0[0x8];
+       u8         tsar_type[0x8];
+       u8         reserved_at_10[0x10];
+};
+
 struct mlx5_ifc_teardown_hca_out_bits {
        u8         status[0x8];
        u8         reserved_at_8[0x18];
@@ -3540,6 +3607,39 @@ struct mlx5_ifc_query_special_contexts_in_bits {
        u8         reserved_at_40[0x40];
 };
 
+struct mlx5_ifc_query_scheduling_element_out_bits {
+       u8         opcode[0x10];
+       u8         reserved_at_10[0x10];
+
+       u8         reserved_at_20[0x10];
+       u8         op_mod[0x10];
+
+       u8         reserved_at_40[0xc0];
+
+       struct mlx5_ifc_scheduling_context_bits scheduling_context;
+
+       u8         reserved_at_300[0x100];
+};
+
+enum {
+       SCHEDULING_HIERARCHY_E_SWITCH = 0x2,
+};
+
+struct mlx5_ifc_query_scheduling_element_in_bits {
+       u8         opcode[0x10];
+       u8         reserved_at_10[0x10];
+
+       u8         reserved_at_20[0x10];
+       u8         op_mod[0x10];
+
+       u8         scheduling_hierarchy[0x8];
+       u8         reserved_at_48[0x18];
+
+       u8         scheduling_element_id[0x20];
+
+       u8         reserved_at_80[0x180];
+};
+
 struct mlx5_ifc_query_rqt_out_bits {
        u8         status[0x8];
        u8         reserved_at_8[0x18];
@@ -4725,6 +4825,43 @@ struct mlx5_ifc_modify_sq_in_bits {
        struct mlx5_ifc_sqc_bits ctx;
 };
 
+struct mlx5_ifc_modify_scheduling_element_out_bits {
+       u8         status[0x8];
+       u8         reserved_at_8[0x18];
+
+       u8         syndrome[0x20];
+
+       u8         reserved_at_40[0x1c0];
+};
+
+enum {
+       MODIFY_SCHEDULING_ELEMENT_IN_MODIFY_BITMASK_BW_SHARE = 0x1,
+       MODIFY_SCHEDULING_ELEMENT_IN_MODIFY_BITMASK_MAX_AVERAGE_BW = 0x2,
+};
+
+struct mlx5_ifc_modify_scheduling_element_in_bits {
+       u8         opcode[0x10];
+       u8         reserved_at_10[0x10];
+
+       u8         reserved_at_20[0x10];
+       u8         op_mod[0x10];
+
+       u8         scheduling_hierarchy[0x8];
+       u8         reserved_at_48[0x18];
+
+       u8         scheduling_element_id[0x20];
+
+       u8         reserved_at_80[0x20];
+
+       u8         modify_bitmask[0x20];
+
+       u8         reserved_at_c0[0x40];
+
+       struct mlx5_ifc_scheduling_context_bits scheduling_context;
+
+       u8         reserved_at_300[0x100];
+};
+
 struct mlx5_ifc_modify_rqt_out_bits {
        u8         status[0x8];
        u8         reserved_at_8[0x18];
@@ -5390,6 +5527,30 @@ struct mlx5_ifc_destroy_sq_in_bits {
        u8         reserved_at_60[0x20];
 };
 
+struct mlx5_ifc_destroy_scheduling_element_out_bits {
+       u8         status[0x8];
+       u8         reserved_at_8[0x18];
+
+       u8         syndrome[0x20];
+
+       u8         reserved_at_40[0x1c0];
+};
+
+struct mlx5_ifc_destroy_scheduling_element_in_bits {
+       u8         opcode[0x10];
+       u8         reserved_at_10[0x10];
+
+       u8         reserved_at_20[0x10];
+       u8         op_mod[0x10];
+
+       u8         scheduling_hierarchy[0x8];
+       u8         reserved_at_48[0x18];
+
+       u8         scheduling_element_id[0x20];
+
+       u8         reserved_at_80[0x180];
+};
+
 struct mlx5_ifc_destroy_rqt_out_bits {
        u8         status[0x8];
        u8         reserved_at_8[0x18];
@@ -6017,6 +6178,36 @@ struct mlx5_ifc_create_sq_in_bits {
        struct mlx5_ifc_sqc_bits ctx;
 };
 
+struct mlx5_ifc_create_scheduling_element_out_bits {
+       u8         status[0x8];
+       u8         reserved_at_8[0x18];
+
+       u8         syndrome[0x20];
+
+       u8         reserved_at_40[0x40];
+
+       u8         scheduling_element_id[0x20];
+
+       u8         reserved_at_a0[0x160];
+};
+
+struct mlx5_ifc_create_scheduling_element_in_bits {
+       u8         opcode[0x10];
+       u8         reserved_at_10[0x10];
+
+       u8         reserved_at_20[0x10];
+       u8         op_mod[0x10];
+
+       u8         scheduling_hierarchy[0x8];
+       u8         reserved_at_48[0x18];
+
+       u8         reserved_at_60[0xa0];
+
+       struct mlx5_ifc_scheduling_context_bits scheduling_context;
+
+       u8         reserved_at_300[0x100];
+};
+
 struct mlx5_ifc_create_rqt_out_bits {
        u8         status[0x8];
        u8         reserved_at_8[0x18];