diff options
author | Kevin Traynor <ktraynor@redhat.com> | 2021-07-16 17:02:09 +0100 |
---|---|---|
committer | Ian Stokes <ian.stokes@intel.com> | 2021-07-16 16:51:47 +0100 |
commit | 3dd050909a74229f024da30ae4800ede79883248 (patch) | |
tree | 129cc8b9426ed20a9552fa7d44a98efee8b7eca5 /Documentation | |
parent | 4fb54652e010c4fc3d4e2d310c64465ffcad2c23 (diff) | |
download | openvswitch-3dd050909a74229f024da30ae4800ede79883248.tar.gz |
dpif-netdev: Add group rxq scheduling assignment type.
Add an rxq scheduling option that allows rxqs to be grouped
on a pmd based purely on their load.
The current default 'cycles' assignment sorts rxqs by measured
processing load and then assigns them to a list of round robin PMDs.
This helps to keep the rxqs that require most processing on different
cores but as it selects the PMDs in round robin order, it equally
distributes rxqs to PMDs.
'cycles' assignment has the advantage in that it separates the most
loaded rxqs from being on the same core but maintains the rxqs being
spread across a broad range of PMDs to mitigate against changes to
traffic pattern.
'cycles' assignment has the disadvantage that in order to make the
trade off between optimising for current traffic load and mitigating
against future changes, it tries to assign and equal amount of rxqs
per PMD in a round robin manner and this can lead to a less than optimal
balance of the processing load.
Now that PMD auto load balance can help mitigate with future changes in
traffic patterns, a 'group' assignment can be used to assign rxqs based
on their measured cycles and the estimated running total of the PMDs.
In this case, there is no restriction about keeping equal number of
rxqs per PMD as it is purely load based.
This means that one PMD may have a group of low load rxqs assigned to it
while another PMD has one high load rxq assigned to it, as that is the
best balance of their measured loads across the PMDs.
Signed-off-by: Kevin Traynor <ktraynor@redhat.com>
Acked-by: Sunil Pai G <sunil.pai.g@intel.com>
Acked-by: David Marchand <david.marchand@redhat.com>
Signed-off-by: Ian Stokes <ian.stokes@intel.com>
Diffstat (limited to 'Documentation')
-rw-r--r-- | Documentation/topics/dpdk/pmd.rst | 26 |
1 files changed, 26 insertions, 0 deletions
diff --git a/Documentation/topics/dpdk/pmd.rst b/Documentation/topics/dpdk/pmd.rst index 065bd16ef..29ba53954 100644 --- a/Documentation/topics/dpdk/pmd.rst +++ b/Documentation/topics/dpdk/pmd.rst @@ -136,6 +136,32 @@ The Rx queues will be assigned to the cores in the following order:: Core 7: Q4 (70%) | Q5 (10%) Core 8: Q3 (60%) | Q0 (30%) +``group`` assignment is similar to ``cycles`` in that the Rxqs will be +ordered by their measured processing cycles before being assigned to PMDs. +It differs from ``cycles`` in that it uses a running estimate of the cycles +that will be on each PMD to select the PMD with the lowest load for each Rxq. + +This means that there can be a group of low traffic Rxqs on one PMD, while a +high traffic Rxq may have a PMD to itself. Where ``cycles`` kept as close to +the same number of Rxqs per PMD as possible, with ``group`` this restriction is +removed for a better balance of the workload across PMDs. + +For example, where there are five Rx queues and three cores - 3, 7, and 8 - +available and the measured usage of core cycles per Rx queue over the last +interval is seen to be: + +- Queue #0: 10% +- Queue #1: 80% +- Queue #3: 50% +- Queue #4: 70% +- Queue #5: 10% + +The Rx queues will be assigned to the cores in the following order:: + + Core 3: Q1 (80%) | + Core 7: Q4 (70%) | + Core 8: Q3 (50%) | Q0 (10%) | Q5 (10%) + Alternatively, ``roundrobin`` assignment can be used, where the Rxqs are assigned to PMDs in a round-robined fashion. This algorithm was used by default prior to OVS 2.9. For example, given the following ports and queues: |