[Intel-wired-lan] [RFC PATCH 15/30] i40e/netpolicy: add MIX policy support

kan.liang at intel.com kan.liang at intel.com
Mon Jul 18 06:56:09 UTC 2016


From: Kan Liang <kan.liang at intel.com>

Enable i40e MIX policy support. Based on the test, the MIX policy has
better performance if increasing rx interrupt moderation a little bit.

For evaluating the MIX policy performance, mixed workloads are tested.
The mixed workloads are combination of throughput-first workload and
latency-first workload. Five different types of combinations are
evaluated.
(pure throughput-first workload, pure latency-first workloads,
 2/3 throughput-first workload + 1/3 latency-first workloads,
 1/3 throughput-first workload + 2/3 latency-first workloads and
 1/2 throughput-first workload + 1/2 latency-first workloads).

For caculating the performance of mixed workloads, a weighted sum system
is also introduced. Here is the formula.

Score = normalized_latency * Weight + normalized_throughput * (1 -
Weight).

If we assume that the user has an equal interest in latency and
throughput performance, the Score for "MIX" policy is on average ~1.52X
than baseline.

Signed-off-by: Kan Liang <kan.liang at intel.com>
---
 drivers/net/ethernet/intel/i40e/i40e_main.c | 71 +++++++++++++++++++++--------
 1 file changed, 51 insertions(+), 20 deletions(-)

diff --git a/drivers/net/ethernet/intel/i40e/i40e_main.c b/drivers/net/ethernet/intel/i40e/i40e_main.c
index 11b921b..d3f087d 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_main.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_main.c
@@ -8966,6 +8966,8 @@ static netdev_features_t i40e_features_check(struct sk_buff *skb,
 #define NET_POLICY_CPU_TX	250
 #define NET_POLICY_BULK_RX	50
 #define NET_POLICY_BULK_TX	125
+#define NET_POLICY_MIX_BULK_RX	62
+#define NET_POLICY_MIX_BULK_TX	122
 #define NET_POLICY_LATENCY_RX	5
 #define NET_POLICY_LATENCY_TX	10
 
@@ -9004,6 +9006,9 @@ static int i40e_ndo_netpolicy_init(struct net_device *dev,
 			set_bit(i, info->avail_policy);
 	}
 
+	/* support MIX policy */
+	info->has_mix_policy = true;
+
 	return 0;
 }
 
@@ -9046,6 +9051,30 @@ static int i40e_ndo_get_irq_info(struct net_device *dev,
 	return 0;
 }
 
+static int i40e_fill_coalesce_for_policy(struct ethtool_coalesce *ec,
+					 enum netpolicy_name name)
+{
+	if (policy_param[name][NETPOLICY_RX] > 0) {
+		ec->rx_coalesce_usecs = policy_param[name][NETPOLICY_RX];
+		ec->use_adaptive_rx_coalesce = 0;
+	} else if (policy_param[name][NETPOLICY_RX] == 0) {
+		ec->use_adaptive_rx_coalesce = 1;
+	} else {
+		return -EINVAL;
+	}
+
+	if (policy_param[name][NETPOLICY_TX] > 0) {
+		ec->tx_coalesce_usecs = policy_param[name][NETPOLICY_TX];
+		ec->use_adaptive_tx_coalesce = 0;
+	} else if (policy_param[name][NETPOLICY_TX] == 0) {
+		ec->use_adaptive_tx_coalesce = 1;
+	} else {
+		return -EINVAL;
+	}
+
+	return 0;
+}
+
 /**
  * i40e_set_net_policy
  * @dev: the net device pointer
@@ -9061,28 +9090,30 @@ static int i40e_set_net_policy(struct net_device *dev,
 	struct i40e_vsi *vsi = np->vsi;
 	struct netpolicy_object *obj;
 	struct ethtool_coalesce ec;
-
-	if (policy_param[name][NETPOLICY_RX] > 0) {
-		ec.rx_coalesce_usecs = policy_param[name][NETPOLICY_RX];
-		ec.use_adaptive_rx_coalesce = 0;
-	} else if (policy_param[name][NETPOLICY_RX] == 0) {
-		ec.use_adaptive_rx_coalesce = 1;
-	} else {
-		return -EINVAL;
-	}
-
-	if (policy_param[name][NETPOLICY_TX] > 0) {
-		ec.tx_coalesce_usecs = policy_param[name][NETPOLICY_TX];
-		ec.use_adaptive_tx_coalesce = 0;
-	} else if (policy_param[name][NETPOLICY_TX] == 0) {
-		ec.use_adaptive_tx_coalesce = 1;
-	} else {
-		return -EINVAL;
-	}
+	int i, ret;
 
 	/*For i40e driver, tx and rx are always in pair */
-	list_for_each_entry(obj, &dev->netpolicy->obj_list[NETPOLICY_RX][name], list) {
-		i40e_set_itr_per_queue(vsi, &ec, obj->queue);
+	if (name == NET_POLICY_MIX) {
+		/* Under MIX policy, the paramers for BULK object are different */
+		policy_param[NET_POLICY_BULK][NETPOLICY_RX] = NET_POLICY_MIX_BULK_RX;
+		policy_param[NET_POLICY_BULK][NETPOLICY_TX] = NET_POLICY_MIX_BULK_TX;
+		for (i = NET_POLICY_NONE; i < NET_POLICY_MAX; i++) {
+			ret = i40e_fill_coalesce_for_policy(&ec, i);
+			if (ret)
+				return ret;
+			list_for_each_entry(obj, &dev->netpolicy->obj_list[NETPOLICY_RX][i], list) {
+				i40e_set_itr_per_queue(vsi, &ec, obj->queue);
+			}
+		}
+	} else {
+		policy_param[NET_POLICY_BULK][NETPOLICY_RX] = NET_POLICY_BULK_RX;
+		policy_param[NET_POLICY_BULK][NETPOLICY_TX] = NET_POLICY_BULK_TX;
+		ret = i40e_fill_coalesce_for_policy(&ec, name);
+		if (ret)
+			return ret;
+		list_for_each_entry(obj, &dev->netpolicy->obj_list[NETPOLICY_RX][name], list) {
+			i40e_set_itr_per_queue(vsi, &ec, obj->queue);
+		}
 	}
 
 	return 0;
-- 
2.5.5



More information about the Intel-wired-lan mailing list