[drm/mgag200] 90f479ae51: vm-scalability.median -18.8% regression

Thomas Zimmermann tzimmermann at suse.de
Sun Aug 4 18:39:19 UTC 2019


Hi

I did some further analysis on this problem and found that the blinking
cursor affects performance of the vm-scalability test case.

I only have a 4-core machine, so scalability is not really testable. Yet
I see the effects of running vm-scalibility against drm-tip, a revert of
the mgag200 patch and the vmap fixes that I posted a few days ago.

After reverting the mgag200 patch, running the test as described in the
report

  bin/lkp run job.yaml

gives results like

  2019-08-02 19:34:37  ./case-anon-cow-seq-hugetlb
  2019-08-02 19:34:37  ./usemem --runtime 300 -n 4 --prealloc --prefault
    -O -U 815395225
  917319627 bytes / 756534 usecs = 1184110 KB/s
  917319627 bytes / 764675 usecs = 1171504 KB/s
  917319627 bytes / 766414 usecs = 1168846 KB/s
  917319627 bytes / 777990 usecs = 1151454 KB/s

Running the test against current drm-tip gives slightly worse results,
such as.

  2019-08-03 19:17:06  ./case-anon-cow-seq-hugetlb
  2019-08-03 19:17:06  ./usemem --runtime 300 -n 4 --prealloc --prefault
    -O -U 815394406
  917318700 bytes / 871607 usecs = 1027778 KB/s
  917318700 bytes / 894173 usecs = 1001840 KB/s
  917318700 bytes / 919694 usecs = 974040 KB/s
  917318700 bytes / 923341 usecs = 970193 KB/s

The test puts out roughly one result per second. Strangely sending the
output to /dev/null can make results significantly worse.

  bin/lkp run job.yaml > /dev/null

  2019-08-03 19:23:04  ./case-anon-cow-seq-hugetlb
  2019-08-03 19:23:04  ./usemem --runtime 300 -n 4 --prealloc --prefault
    -O -U 815394406
  917318700 bytes / 1207358 usecs = 741966 KB/s
  917318700 bytes / 1210456 usecs = 740067 KB/s
  917318700 bytes / 1216572 usecs = 736346 KB/s
  917318700 bytes / 1239152 usecs = 722929 KB/s

I realized that there's still a blinking cursor on the screen, which I
disabled with

  tput civis

or alternatively

  echo 0 > /sys/devices/virtual/graphics/fbcon/cursor_blink

Running the the test now gives the original or even better results, such as

  bin/lkp run job.yaml > /dev/null

  2019-08-03 19:29:17  ./case-anon-cow-seq-hugetlb
  2019-08-03 19:29:17  ./usemem --runtime 300 -n 4 --prealloc --prefault
    -O -U 815394406
  917318700 bytes / 659419 usecs = 1358497 KB/s
  917318700 bytes / 659658 usecs = 1358005 KB/s
  917318700 bytes / 659916 usecs = 1357474 KB/s
  917318700 bytes / 660168 usecs = 1356956 KB/s

Rong, Feng, could you confirm this by disabling the cursor or blinking?


The difference between mgag200's original fbdev support and generic
fbdev emulation is generic fbdev's worker task that updates the VRAM
buffer from the shadow buffer. mgag200 does this immediately, but relies
on drm_can_sleep(), which is deprecated.

I think that the worker task interferes with the test case, as the
worker has been in fbdev emulation since forever and no performance
regressions have been reported so far.


So unless there's a report where this problem happens in a real-world
use case, I'd like to keep code as it is. And apparently there's always
the workaround of disabling the cursor blinking.

Best regards
Thomas


Am 30.07.19 um 19:50 schrieb Thomas Zimmermann:
> Am 29.07.19 um 11:51 schrieb kernel test robot:
>> Greeting,
>>
>> FYI, we noticed a -18.8% regression of vm-scalability.median due to commit:>
>>
>> commit: 90f479ae51afa45efab97afdde9b94b9660dd3e4 ("drm/mgag200: Replace struct mga_fbdev with generic framebuffer emulation")
>> https://kernel.googlesource.com/pub/scm/linux/kernel/git/next/linux-next.git master
> 
> Daniel, Noralf, we may have to revert this patch.
> 
> I expected some change in display performance, but not in VM. Since it's
> a server chipset, probably no one cares much about display performance.
> So that seemed like a good trade-off for re-using shared code.
> 
> Part of the patch set is that the generic fb emulation now maps and
> unmaps the fbdev BO when updating the screen. I guess that's the cause
> of the performance regression. And it should be visible with other
> drivers as well if they use a shadow FB for fbdev emulation.
> 
> The thing is that we'd need another generic fbdev emulation for ast and
> mgag200 that handles this issue properly.
> 
> Best regards
> Thomas
> 
>>
>> in testcase: vm-scalability
>> on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
>> with following parameters:
>>
>> 	runtime: 300s
>> 	size: 8T
>> 	test: anon-cow-seq-hugetlb
>> 	cpufreq_governor: performance
>>
>> test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
>> test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/
>>
>>
>>
>> Details are as below:
>> -------------------------------------------------------------------------------------------------->
>>
>>
>> To reproduce:
>>
>>         git clone https://github.com/intel/lkp-tests.git
>>         cd lkp-tests
>>         bin/lkp install job.yaml  # job file is attached in this email
>>         bin/lkp run     job.yaml
>>
>> =========================================================================================
>> compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
>>   gcc-7/performance/x86_64-rhel-7.6/debian-x86_64-2019-05-14.cgz/300s/8T/lkp-knm01/anon-cow-seq-hugetlb/vm-scalability
>>
>> commit: 
>>   f1f8555dfb ("drm/bochs: Use shadow buffer for bochs framebuffer console")
>>   90f479ae51 ("drm/mgag200: Replace struct mga_fbdev with generic framebuffer emulation")
>>
>> f1f8555dfb9a70a2 90f479ae51afa45efab97afdde9 
>> ---------------- --------------------------- 
>>        fail:runs  %reproduction    fail:runs
>>            |             |             |    
>>           2:4          -50%            :4     dmesg.WARNING:at#for_ip_interrupt_entry/0x
>>            :4           25%           1:4     dmesg.WARNING:at_ip___perf_sw_event/0x
>>            :4           25%           1:4     dmesg.WARNING:at_ip__fsnotify_parent/0x
>>          %stddev     %change         %stddev
>>              \          |                \  
>>      43955 ±  2%     -18.8%      35691        vm-scalability.median
>>       0.06 ±  7%    +193.0%       0.16 ±  2%  vm-scalability.median_stddev
>>   14906559 ±  2%     -17.9%   12237079        vm-scalability.throughput
>>      87651 ±  2%     -17.4%      72374        vm-scalability.time.involuntary_context_switches
>>    2086168           -23.6%    1594224        vm-scalability.time.minor_page_faults
>>      15082 ±  2%     -10.4%      13517        vm-scalability.time.percent_of_cpu_this_job_got
>>      29987            -8.9%      27327        vm-scalability.time.system_time
>>      15755           -12.4%      13795        vm-scalability.time.user_time
>>     122011           -19.3%      98418        vm-scalability.time.voluntary_context_switches
>>  3.034e+09           -23.6%  2.318e+09        vm-scalability.workload
>>     242478 ± 12%     +68.5%     408518 ± 23%  cpuidle.POLL.time
>>       2788 ± 21%    +117.4%       6062 ± 26%  cpuidle.POLL.usage
>>      56653 ± 10%     +64.4%      93144 ± 20%  meminfo.Mapped
>>     120392 ±  7%     +14.0%     137212 ±  4%  meminfo.Shmem
>>      47221 ± 11%     +77.1%      83634 ± 22%  numa-meminfo.node0.Mapped
>>     120465 ±  7%     +13.9%     137205 ±  4%  numa-meminfo.node0.Shmem
>>    2885513           -16.5%    2409384        numa-numastat.node0.local_node
>>    2885471           -16.5%    2409354        numa-numastat.node0.numa_hit
>>      11813 ± 11%     +76.3%      20824 ± 22%  numa-vmstat.node0.nr_mapped
>>      30096 ±  7%     +13.8%      34238 ±  4%  numa-vmstat.node0.nr_shmem
>>      43.72 ±  2%      +5.5       49.20        mpstat.cpu.all.idle%
>>       0.03 ±  4%      +0.0        0.05 ±  6%  mpstat.cpu.all.soft%
>>      19.51            -2.4       17.08        mpstat.cpu.all.usr%
>>       1012            -7.9%     932.75        turbostat.Avg_MHz
>>      32.38 ± 10%     +25.8%      40.73        turbostat.CPU%c1
>>     145.51            -3.1%     141.01        turbostat.PkgWatt
>>      15.09           -19.2%      12.19        turbostat.RAMWatt
>>      43.50 ±  2%     +13.2%      49.25        vmstat.cpu.id
>>      18.75 ±  2%     -13.3%      16.25 ±  2%  vmstat.cpu.us
>>     152.00 ±  2%      -9.5%     137.50        vmstat.procs.r
>>       4800           -13.1%       4173        vmstat.system.cs
>>     156170           -11.9%     137594        slabinfo.anon_vma.active_objs
>>       3395           -11.9%       2991        slabinfo.anon_vma.active_slabs
>>     156190           -11.9%     137606        slabinfo.anon_vma.num_objs
>>       3395           -11.9%       2991        slabinfo.anon_vma.num_slabs
>>       1716 ±  5%     +11.5%       1913 ±  8%  slabinfo.dmaengine-unmap-16.active_objs
>>       1716 ±  5%     +11.5%       1913 ±  8%  slabinfo.dmaengine-unmap-16.num_objs
>>       1767 ±  2%     -19.0%       1431 ±  2%  slabinfo.hugetlbfs_inode_cache.active_objs
>>       1767 ±  2%     -19.0%       1431 ±  2%  slabinfo.hugetlbfs_inode_cache.num_objs
>>       3597 ±  5%     -16.4%       3006 ±  3%  slabinfo.skbuff_ext_cache.active_objs
>>       3597 ±  5%     -16.4%       3006 ±  3%  slabinfo.skbuff_ext_cache.num_objs
>>    1330122           -23.6%    1016557        proc-vmstat.htlb_buddy_alloc_success
>>      77214 ±  3%      +6.4%      82128 ±  2%  proc-vmstat.nr_active_anon
>>      67277            +2.9%      69246        proc-vmstat.nr_anon_pages
>>     218.50 ±  3%     -10.6%     195.25        proc-vmstat.nr_dirtied
>>     288628            +1.4%     292755        proc-vmstat.nr_file_pages
>>     360.50            -2.7%     350.75        proc-vmstat.nr_inactive_file
>>      14225 ±  9%     +63.8%      23304 ± 20%  proc-vmstat.nr_mapped
>>      30109 ±  7%     +13.8%      34259 ±  4%  proc-vmstat.nr_shmem
>>      99870            -1.3%      98597        proc-vmstat.nr_slab_unreclaimable
>>     204.00 ±  4%     -12.1%     179.25        proc-vmstat.nr_written
>>      77214 ±  3%      +6.4%      82128 ±  2%  proc-vmstat.nr_zone_active_anon
>>     360.50            -2.7%     350.75        proc-vmstat.nr_zone_inactive_file
>>       8810 ± 19%     -66.1%       2987 ± 42%  proc-vmstat.numa_hint_faults
>>       8810 ± 19%     -66.1%       2987 ± 42%  proc-vmstat.numa_hint_faults_local
>>    2904082           -16.4%    2427026        proc-vmstat.numa_hit
>>    2904081           -16.4%    2427025        proc-vmstat.numa_local
>>  6.828e+08           -23.5%  5.221e+08        proc-vmstat.pgalloc_normal
>>    2900008           -17.2%    2400195        proc-vmstat.pgfault
>>  6.827e+08           -23.5%   5.22e+08        proc-vmstat.pgfree
>>  1.635e+10           -17.0%  1.357e+10        perf-stat.i.branch-instructions
>>       1.53 ±  4%      -0.1        1.45 ±  3%  perf-stat.i.branch-miss-rate%
>>  2.581e+08 ±  3%     -20.5%  2.051e+08 ±  2%  perf-stat.i.branch-misses
>>      12.66            +1.1       13.78        perf-stat.i.cache-miss-rate%
>>   72720849           -12.0%   63958986        perf-stat.i.cache-misses
>>  5.766e+08           -18.6%  4.691e+08        perf-stat.i.cache-references
>>       4674 ±  2%     -13.0%       4064        perf-stat.i.context-switches
>>       4.29           +12.5%       4.83        perf-stat.i.cpi
>>  2.573e+11            -7.4%  2.383e+11        perf-stat.i.cpu-cycles
>>     231.35           -21.5%     181.56        perf-stat.i.cpu-migrations
>>       3522            +4.4%       3677        perf-stat.i.cycles-between-cache-misses
>>       0.09 ± 13%      +0.0        0.12 ±  5%  perf-stat.i.iTLB-load-miss-rate%
>>  5.894e+10           -15.8%  4.961e+10        perf-stat.i.iTLB-loads
>>  5.901e+10           -15.8%  4.967e+10        perf-stat.i.instructions
>>       1291 ± 14%     -21.8%       1010        perf-stat.i.instructions-per-iTLB-miss
>>       0.24           -11.0%       0.21        perf-stat.i.ipc
>>       9476           -17.5%       7821        perf-stat.i.minor-faults
>>       9478           -17.5%       7821        perf-stat.i.page-faults
>>       9.76            -3.6%       9.41        perf-stat.overall.MPKI
>>       1.59 ±  4%      -0.1        1.52        perf-stat.overall.branch-miss-rate%
>>      12.61            +1.1       13.71        perf-stat.overall.cache-miss-rate%
>>       4.38           +10.5%       4.83        perf-stat.overall.cpi
>>       3557            +5.3%       3747        perf-stat.overall.cycles-between-cache-misses
>>       0.08 ± 12%      +0.0        0.10        perf-stat.overall.iTLB-load-miss-rate%
>>       1268 ± 15%     -23.0%     976.22        perf-stat.overall.instructions-per-iTLB-miss
>>       0.23            -9.5%       0.21        perf-stat.overall.ipc
>>       5815            +9.7%       6378        perf-stat.overall.path-length
>>  1.634e+10           -17.5%  1.348e+10        perf-stat.ps.branch-instructions
>>  2.595e+08 ±  3%     -21.2%  2.043e+08 ±  2%  perf-stat.ps.branch-misses
>>   72565205           -12.2%   63706339        perf-stat.ps.cache-misses
>>  5.754e+08           -19.2%  4.646e+08        perf-stat.ps.cache-references
>>       4640 ±  2%     -12.5%       4060        perf-stat.ps.context-switches
>>  2.581e+11            -7.5%  2.387e+11        perf-stat.ps.cpu-cycles
>>     229.91           -22.0%     179.42        perf-stat.ps.cpu-migrations
>>  5.889e+10           -16.3%  4.927e+10        perf-stat.ps.iTLB-loads
>>  5.899e+10           -16.3%  4.938e+10        perf-stat.ps.instructions
>>       9388           -18.2%       7677        perf-stat.ps.minor-faults
>>       9389           -18.2%       7677        perf-stat.ps.page-faults
>>  1.764e+13           -16.2%  1.479e+13        perf-stat.total.instructions
>>      46803 ±  3%     -18.8%      37982 ±  6%  sched_debug.cfs_rq:/.exec_clock.min
>>       5320 ±  3%     +23.7%       6581 ±  3%  sched_debug.cfs_rq:/.exec_clock.stddev
>>       6737 ± 14%     +58.1%      10649 ± 10%  sched_debug.cfs_rq:/.load.avg
>>     587978 ± 17%     +58.2%     930382 ±  9%  sched_debug.cfs_rq:/.load.max
>>      46952 ± 16%     +64.8%      77388 ± 11%  sched_debug.cfs_rq:/.load.stddev
>>       7.12 ±  4%     +49.1%      10.62 ±  6%  sched_debug.cfs_rq:/.load_avg.avg
>>     474.40 ± 23%     +67.5%     794.60 ± 10%  sched_debug.cfs_rq:/.load_avg.max
>>      37.70 ± 11%     +74.8%      65.90 ±  9%  sched_debug.cfs_rq:/.load_avg.stddev
>>   13424269 ±  4%     -15.6%   11328098 ±  2%  sched_debug.cfs_rq:/.min_vruntime.avg
>>   15411275 ±  3%     -12.4%   13505072 ±  2%  sched_debug.cfs_rq:/.min_vruntime.max
>>    7939295 ±  6%     -17.5%    6551322 ±  7%  sched_debug.cfs_rq:/.min_vruntime.min
>>      21.44 ±  7%     -56.1%       9.42 ±  4%  sched_debug.cfs_rq:/.nr_spread_over.avg
>>     117.45 ± 11%     -60.6%      46.30 ± 14%  sched_debug.cfs_rq:/.nr_spread_over.max
>>      19.33 ±  8%     -66.4%       6.49 ±  9%  sched_debug.cfs_rq:/.nr_spread_over.stddev
>>       4.32 ± 15%     +84.4%       7.97 ±  3%  sched_debug.cfs_rq:/.runnable_load_avg.avg
>>     353.85 ± 29%    +118.8%     774.35 ± 11%  sched_debug.cfs_rq:/.runnable_load_avg.max
>>      27.30 ± 24%    +118.5%      59.64 ±  9%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
>>       6729 ± 14%     +58.2%      10644 ± 10%  sched_debug.cfs_rq:/.runnable_weight.avg
>>     587978 ± 17%     +58.2%     930382 ±  9%  sched_debug.cfs_rq:/.runnable_weight.max
>>      46950 ± 16%     +64.8%      77387 ± 11%  sched_debug.cfs_rq:/.runnable_weight.stddev
>>    5305069 ±  4%     -17.4%    4380376 ±  7%  sched_debug.cfs_rq:/.spread0.avg
>>    7328745 ±  3%      -9.9%    6600897 ±  3%  sched_debug.cfs_rq:/.spread0.max
>>    2220837 ±  4%     +55.8%    3460596 ±  5%  sched_debug.cpu.avg_idle.avg
>>    4590666 ±  9%     +76.8%    8117037 ± 15%  sched_debug.cpu.avg_idle.max
>>     485052 ±  7%     +80.3%     874679 ± 10%  sched_debug.cpu.avg_idle.stddev
>>     561.50 ± 26%     +37.7%     773.30 ± 15%  sched_debug.cpu.clock.stddev
>>     561.50 ± 26%     +37.7%     773.30 ± 15%  sched_debug.cpu.clock_task.stddev
>>       3.20 ± 10%    +109.6%       6.70 ±  3%  sched_debug.cpu.cpu_load[0].avg
>>     309.10 ± 20%    +150.3%     773.75 ± 12%  sched_debug.cpu.cpu_load[0].max
>>      21.02 ± 14%    +160.8%      54.80 ±  9%  sched_debug.cpu.cpu_load[0].stddev
>>       3.19 ±  8%    +109.8%       6.70 ±  3%  sched_debug.cpu.cpu_load[1].avg
>>     299.75 ± 19%    +158.0%     773.30 ± 12%  sched_debug.cpu.cpu_load[1].max
>>      20.32 ± 12%    +168.7%      54.62 ±  9%  sched_debug.cpu.cpu_load[1].stddev
>>       3.20 ±  8%    +109.1%       6.69 ±  4%  sched_debug.cpu.cpu_load[2].avg
>>     288.90 ± 20%    +167.0%     771.40 ± 12%  sched_debug.cpu.cpu_load[2].max
>>      19.70 ± 12%    +175.4%      54.27 ±  9%  sched_debug.cpu.cpu_load[2].stddev
>>       3.16 ±  8%    +110.9%       6.66 ±  6%  sched_debug.cpu.cpu_load[3].avg
>>     275.50 ± 24%    +178.4%     766.95 ± 12%  sched_debug.cpu.cpu_load[3].max
>>      18.92 ± 15%    +184.2%      53.77 ± 10%  sched_debug.cpu.cpu_load[3].stddev
>>       3.08 ±  8%    +115.7%       6.65 ±  7%  sched_debug.cpu.cpu_load[4].avg
>>     263.55 ± 28%    +188.7%     760.85 ± 12%  sched_debug.cpu.cpu_load[4].max
>>      18.03 ± 18%    +196.6%      53.46 ± 11%  sched_debug.cpu.cpu_load[4].stddev
>>      14543            -9.6%      13150        sched_debug.cpu.curr->pid.max
>>       5293 ± 16%     +74.7%       9248 ± 11%  sched_debug.cpu.load.avg
>>     587978 ± 17%     +58.2%     930382 ±  9%  sched_debug.cpu.load.max
>>      40887 ± 19%     +78.3%      72891 ±  9%  sched_debug.cpu.load.stddev
>>    1141679 ±  4%     +56.9%    1790907 ±  5%  sched_debug.cpu.max_idle_balance_cost.avg
>>    2432100 ±  9%     +72.6%    4196779 ± 13%  sched_debug.cpu.max_idle_balance_cost.max
>>     745656           +29.3%     964170 ±  5%  sched_debug.cpu.max_idle_balance_cost.min
>>     239032 ±  9%     +81.9%     434806 ± 10%  sched_debug.cpu.max_idle_balance_cost.stddev
>>       0.00 ± 27%     +92.1%       0.00 ± 31%  sched_debug.cpu.next_balance.stddev
>>       1030 ±  4%     -10.4%     924.00 ±  2%  sched_debug.cpu.nr_switches.min
>>       0.04 ± 26%    +139.0%       0.09 ± 41%  sched_debug.cpu.nr_uninterruptible.avg
>>     830.35 ±  6%     -12.0%     730.50 ±  2%  sched_debug.cpu.sched_count.min
>>     912.00 ±  2%      -9.5%     825.38        sched_debug.cpu.ttwu_count.avg
>>     433.05 ±  3%     -19.2%     350.05 ±  3%  sched_debug.cpu.ttwu_count.min
>>     160.70 ±  3%     -12.5%     140.60 ±  4%  sched_debug.cpu.ttwu_local.min
>>       9072 ± 11%     -36.4%       5767 ±  8%  softirqs.CPU1.RCU
>>      12769 ±  5%     +15.3%      14718 ±  3%  softirqs.CPU101.SCHED
>>      13198           +11.5%      14717 ±  3%  softirqs.CPU102.SCHED
>>      12981 ±  4%     +13.9%      14788 ±  3%  softirqs.CPU105.SCHED
>>      13486 ±  3%     +11.8%      15071 ±  4%  softirqs.CPU111.SCHED
>>      12794 ±  4%     +14.1%      14601 ±  9%  softirqs.CPU112.SCHED
>>      12999 ±  4%     +10.1%      14314 ±  4%  softirqs.CPU115.SCHED
>>      12844 ±  4%     +10.6%      14202 ±  2%  softirqs.CPU120.SCHED
>>      13336 ±  3%      +9.4%      14585 ±  3%  softirqs.CPU122.SCHED
>>      12639 ±  4%     +20.2%      15195        softirqs.CPU123.SCHED
>>      13040 ±  5%     +15.2%      15024 ±  5%  softirqs.CPU126.SCHED
>>      13123           +15.1%      15106 ±  5%  softirqs.CPU127.SCHED
>>       9188 ±  6%     -35.7%       5911 ±  2%  softirqs.CPU13.RCU
>>      13054 ±  3%     +13.1%      14761 ±  5%  softirqs.CPU130.SCHED
>>      13158 ±  2%     +13.9%      14985 ±  5%  softirqs.CPU131.SCHED
>>      12797 ±  6%     +13.5%      14524 ±  3%  softirqs.CPU133.SCHED
>>      12452 ±  5%     +14.8%      14297        softirqs.CPU134.SCHED
>>      13078 ±  3%     +10.4%      14439 ±  3%  softirqs.CPU138.SCHED
>>      12617 ±  2%     +14.5%      14442 ±  5%  softirqs.CPU139.SCHED
>>      12974 ±  3%     +13.7%      14752 ±  4%  softirqs.CPU142.SCHED
>>      12579 ±  4%     +19.1%      14983 ±  3%  softirqs.CPU143.SCHED
>>       9122 ± 24%     -44.6%       5053 ±  5%  softirqs.CPU144.RCU
>>      13366 ±  2%     +11.1%      14848 ±  3%  softirqs.CPU149.SCHED
>>      13246 ±  2%     +22.0%      16162 ±  7%  softirqs.CPU150.SCHED
>>      13452 ±  3%     +20.5%      16210 ±  7%  softirqs.CPU151.SCHED
>>      13507           +10.1%      14869        softirqs.CPU156.SCHED
>>      13808 ±  3%      +9.2%      15079 ±  4%  softirqs.CPU157.SCHED
>>      13442 ±  2%     +13.4%      15248 ±  4%  softirqs.CPU160.SCHED
>>      13311           +12.1%      14920 ±  2%  softirqs.CPU162.SCHED
>>      13544 ±  3%      +8.5%      14695 ±  4%  softirqs.CPU163.SCHED
>>      13648 ±  3%     +11.2%      15179 ±  2%  softirqs.CPU166.SCHED
>>      13404 ±  4%     +12.5%      15079 ±  3%  softirqs.CPU168.SCHED
>>      13421 ±  6%     +16.0%      15568 ±  8%  softirqs.CPU169.SCHED
>>      13115 ±  3%     +23.1%      16139 ± 10%  softirqs.CPU171.SCHED
>>      13424 ±  6%     +10.4%      14822 ±  3%  softirqs.CPU175.SCHED
>>      13274 ±  3%     +13.7%      15087 ±  9%  softirqs.CPU185.SCHED
>>      13409 ±  3%     +12.3%      15063 ±  3%  softirqs.CPU190.SCHED
>>      13181 ±  7%     +13.4%      14946 ±  3%  softirqs.CPU196.SCHED
>>      13578 ±  3%     +10.9%      15061        softirqs.CPU197.SCHED
>>      13323 ±  5%     +24.8%      16627 ±  6%  softirqs.CPU198.SCHED
>>      14072 ±  2%     +12.3%      15798 ±  7%  softirqs.CPU199.SCHED
>>      12604 ± 13%     +17.9%      14865        softirqs.CPU201.SCHED
>>      13380 ±  4%     +14.8%      15356 ±  3%  softirqs.CPU203.SCHED
>>      13481 ±  8%     +14.2%      15390 ±  3%  softirqs.CPU204.SCHED
>>      12921 ±  2%     +13.8%      14710 ±  3%  softirqs.CPU206.SCHED
>>      13468           +13.0%      15218 ±  2%  softirqs.CPU208.SCHED
>>      13253 ±  2%     +13.1%      14992        softirqs.CPU209.SCHED
>>      13319 ±  2%     +14.3%      15225 ±  7%  softirqs.CPU210.SCHED
>>      13673 ±  5%     +16.3%      15895 ±  3%  softirqs.CPU211.SCHED
>>      13290           +17.0%      15556 ±  5%  softirqs.CPU212.SCHED
>>      13455 ±  4%     +14.4%      15392 ±  3%  softirqs.CPU213.SCHED
>>      13454 ±  4%     +14.3%      15377 ±  3%  softirqs.CPU215.SCHED
>>      13872 ±  7%      +9.7%      15221 ±  5%  softirqs.CPU220.SCHED
>>      13555 ±  4%     +17.3%      15896 ±  5%  softirqs.CPU222.SCHED
>>      13411 ±  4%     +20.8%      16197 ±  6%  softirqs.CPU223.SCHED
>>       8472 ± 21%     -44.8%       4680 ±  3%  softirqs.CPU224.RCU
>>      13141 ±  3%     +16.2%      15265 ±  7%  softirqs.CPU225.SCHED
>>      14084 ±  3%      +8.2%      15242 ±  2%  softirqs.CPU226.SCHED
>>      13528 ±  4%     +11.3%      15063 ±  4%  softirqs.CPU228.SCHED
>>      13218 ±  3%     +16.3%      15377 ±  4%  softirqs.CPU229.SCHED
>>      14031 ±  4%     +10.2%      15467 ±  2%  softirqs.CPU231.SCHED
>>      13770 ±  3%     +14.0%      15700 ±  3%  softirqs.CPU232.SCHED
>>      13456 ±  3%     +12.3%      15105 ±  3%  softirqs.CPU233.SCHED
>>      13137 ±  4%     +13.5%      14909 ±  3%  softirqs.CPU234.SCHED
>>      13318 ±  2%     +14.7%      15280 ±  2%  softirqs.CPU235.SCHED
>>      13690 ±  2%     +13.7%      15563 ±  7%  softirqs.CPU238.SCHED
>>      13771 ±  5%     +20.8%      16634 ±  7%  softirqs.CPU241.SCHED
>>      13317 ±  7%     +19.5%      15919 ±  9%  softirqs.CPU243.SCHED
>>       8234 ± 16%     -43.9%       4616 ±  5%  softirqs.CPU244.RCU
>>      13845 ±  6%     +13.0%      15643 ±  3%  softirqs.CPU244.SCHED
>>      13179 ±  3%     +16.3%      15323        softirqs.CPU246.SCHED
>>      13754           +12.2%      15438 ±  3%  softirqs.CPU248.SCHED
>>      13769 ±  4%     +10.9%      15276 ±  2%  softirqs.CPU252.SCHED
>>      13702           +10.5%      15147 ±  2%  softirqs.CPU254.SCHED
>>      13315 ±  2%     +12.5%      14980 ±  3%  softirqs.CPU255.SCHED
>>      13785 ±  3%     +12.9%      15568 ±  5%  softirqs.CPU256.SCHED
>>      13307 ±  3%     +15.0%      15298 ±  3%  softirqs.CPU257.SCHED
>>      13864 ±  3%     +10.5%      15313 ±  2%  softirqs.CPU259.SCHED
>>      13879 ±  2%     +11.4%      15465        softirqs.CPU261.SCHED
>>      13815           +13.6%      15687 ±  5%  softirqs.CPU264.SCHED
>>     119574 ±  2%     +11.8%     133693 ± 11%  softirqs.CPU266.TIMER
>>      13688           +10.9%      15180 ±  6%  softirqs.CPU267.SCHED
>>      11716 ±  4%     +19.3%      13974 ±  8%  softirqs.CPU27.SCHED
>>      13866 ±  3%     +13.7%      15765 ±  4%  softirqs.CPU271.SCHED
>>      13887 ±  5%     +12.5%      15621        softirqs.CPU272.SCHED
>>      13383 ±  3%     +19.8%      16031 ±  2%  softirqs.CPU274.SCHED
>>      13347           +14.1%      15232 ±  3%  softirqs.CPU275.SCHED
>>      12884 ±  2%     +21.0%      15593 ±  4%  softirqs.CPU276.SCHED
>>      13131 ±  5%     +13.4%      14891 ±  5%  softirqs.CPU277.SCHED
>>      12891 ±  2%     +19.2%      15371 ±  4%  softirqs.CPU278.SCHED
>>      13313 ±  4%     +13.0%      15049 ±  2%  softirqs.CPU279.SCHED
>>      13514 ±  3%     +10.2%      14897 ±  2%  softirqs.CPU280.SCHED
>>      13501 ±  3%     +13.7%      15346        softirqs.CPU281.SCHED
>>      13261           +17.5%      15577        softirqs.CPU282.SCHED
>>       8076 ± 15%     -43.7%       4546 ±  5%  softirqs.CPU283.RCU
>>      13686 ±  3%     +12.6%      15413 ±  2%  softirqs.CPU284.SCHED
>>      13439 ±  2%      +9.2%      14670 ±  4%  softirqs.CPU285.SCHED
>>       8878 ±  9%     -35.4%       5735 ±  4%  softirqs.CPU35.RCU
>>      11690 ±  2%     +13.6%      13274 ±  5%  softirqs.CPU40.SCHED
>>      11714 ±  2%     +19.3%      13975 ± 13%  softirqs.CPU41.SCHED
>>      11763           +12.5%      13239 ±  4%  softirqs.CPU45.SCHED
>>      11662 ±  2%      +9.4%      12757 ±  3%  softirqs.CPU46.SCHED
>>      11805 ±  2%      +9.3%      12902 ±  2%  softirqs.CPU50.SCHED
>>      12158 ±  3%     +12.3%      13655 ±  8%  softirqs.CPU55.SCHED
>>      11716 ±  4%      +8.8%      12751 ±  3%  softirqs.CPU58.SCHED
>>      11922 ±  2%      +9.9%      13100 ±  4%  softirqs.CPU64.SCHED
>>       9674 ± 17%     -41.8%       5625 ±  6%  softirqs.CPU66.RCU
>>      11818           +12.0%      13237        softirqs.CPU66.SCHED
>>     124682 ±  7%      -6.1%     117088 ±  5%  softirqs.CPU66.TIMER
>>       8637 ±  9%     -34.0%       5700 ±  7%  softirqs.CPU70.RCU
>>      11624 ±  2%     +11.0%      12901 ±  2%  softirqs.CPU70.SCHED
>>      12372 ±  2%     +13.2%      14003 ±  3%  softirqs.CPU71.SCHED
>>       9949 ± 25%     -33.9%       6574 ± 31%  softirqs.CPU72.RCU
>>      10392 ± 26%     -35.1%       6745 ± 35%  softirqs.CPU73.RCU
>>      12766 ±  3%     +11.1%      14188 ±  3%  softirqs.CPU76.SCHED
>>      12611 ±  2%     +18.8%      14984 ±  5%  softirqs.CPU78.SCHED
>>      12786 ±  3%     +17.9%      15079 ±  7%  softirqs.CPU79.SCHED
>>      11947 ±  4%      +9.7%      13103 ±  4%  softirqs.CPU8.SCHED
>>      13379 ±  7%     +11.8%      14962 ±  4%  softirqs.CPU83.SCHED
>>      13438 ±  5%      +9.7%      14738 ±  2%  softirqs.CPU84.SCHED
>>      12768           +19.4%      15241 ±  6%  softirqs.CPU88.SCHED
>>       8604 ± 13%     -39.3%       5222 ±  3%  softirqs.CPU89.RCU
>>      13077 ±  2%     +17.1%      15308 ±  7%  softirqs.CPU89.SCHED
>>      11887 ±  3%     +20.1%      14272 ±  5%  softirqs.CPU9.SCHED
>>      12723 ±  3%     +11.3%      14165 ±  4%  softirqs.CPU90.SCHED
>>       8439 ± 12%     -38.9%       5153 ±  4%  softirqs.CPU91.RCU
>>      13429 ±  3%     +10.3%      14806 ±  2%  softirqs.CPU95.SCHED
>>      12852 ±  4%     +10.3%      14174 ±  5%  softirqs.CPU96.SCHED
>>      13010 ±  2%     +14.4%      14888 ±  5%  softirqs.CPU97.SCHED
>>    2315644 ±  4%     -36.2%    1477200 ±  4%  softirqs.RCU
>>       1572 ± 10%     +63.9%       2578 ± 39%  interrupts.CPU0.NMI:Non-maskable_interrupts
>>       1572 ± 10%     +63.9%       2578 ± 39%  interrupts.CPU0.PMI:Performance_monitoring_interrupts
>>     252.00 ± 11%     -35.2%     163.25 ± 13%  interrupts.CPU104.RES:Rescheduling_interrupts
>>       2738 ± 24%     +52.4%       4173 ± 19%  interrupts.CPU105.NMI:Non-maskable_interrupts
>>       2738 ± 24%     +52.4%       4173 ± 19%  interrupts.CPU105.PMI:Performance_monitoring_interrupts
>>     245.75 ± 19%     -31.0%     169.50 ±  7%  interrupts.CPU105.RES:Rescheduling_interrupts
>>     228.75 ± 13%     -24.7%     172.25 ± 19%  interrupts.CPU106.RES:Rescheduling_interrupts
>>       2243 ± 15%     +66.3%       3730 ± 35%  interrupts.CPU113.NMI:Non-maskable_interrupts
>>       2243 ± 15%     +66.3%       3730 ± 35%  interrupts.CPU113.PMI:Performance_monitoring_interrupts
>>       2703 ± 31%     +67.0%       4514 ± 33%  interrupts.CPU118.NMI:Non-maskable_interrupts
>>       2703 ± 31%     +67.0%       4514 ± 33%  interrupts.CPU118.PMI:Performance_monitoring_interrupts
>>       2613 ± 25%     +42.2%       3715 ± 24%  interrupts.CPU121.NMI:Non-maskable_interrupts
>>       2613 ± 25%     +42.2%       3715 ± 24%  interrupts.CPU121.PMI:Performance_monitoring_interrupts
>>     311.50 ± 23%     -47.7%     163.00 ±  9%  interrupts.CPU122.RES:Rescheduling_interrupts
>>     266.75 ± 19%     -31.6%     182.50 ± 15%  interrupts.CPU124.RES:Rescheduling_interrupts
>>     293.75 ± 33%     -32.3%     198.75 ± 19%  interrupts.CPU125.RES:Rescheduling_interrupts
>>       2601 ± 36%     +43.2%       3724 ± 29%  interrupts.CPU127.NMI:Non-maskable_interrupts
>>       2601 ± 36%     +43.2%       3724 ± 29%  interrupts.CPU127.PMI:Performance_monitoring_interrupts
>>       2258 ± 21%     +68.2%       3797 ± 29%  interrupts.CPU13.NMI:Non-maskable_interrupts
>>       2258 ± 21%     +68.2%       3797 ± 29%  interrupts.CPU13.PMI:Performance_monitoring_interrupts
>>       3338 ± 29%     +54.6%       5160 ±  9%  interrupts.CPU139.NMI:Non-maskable_interrupts
>>       3338 ± 29%     +54.6%       5160 ±  9%  interrupts.CPU139.PMI:Performance_monitoring_interrupts
>>     219.50 ± 27%     -23.0%     169.00 ± 21%  interrupts.CPU139.RES:Rescheduling_interrupts
>>     290.25 ± 25%     -32.5%     196.00 ± 11%  interrupts.CPU14.RES:Rescheduling_interrupts
>>     243.50 ±  4%     -16.0%     204.50 ± 12%  interrupts.CPU140.RES:Rescheduling_interrupts
>>       1797 ± 15%    +135.0%       4223 ± 46%  interrupts.CPU147.NMI:Non-maskable_interrupts
>>       1797 ± 15%    +135.0%       4223 ± 46%  interrupts.CPU147.PMI:Performance_monitoring_interrupts
>>       2537 ± 22%     +89.6%       4812 ± 28%  interrupts.CPU15.NMI:Non-maskable_interrupts
>>       2537 ± 22%     +89.6%       4812 ± 28%  interrupts.CPU15.PMI:Performance_monitoring_interrupts
>>     292.25 ± 34%     -33.9%     193.25 ±  6%  interrupts.CPU15.RES:Rescheduling_interrupts
>>     424.25 ± 37%     -58.5%     176.25 ± 14%  interrupts.CPU158.RES:Rescheduling_interrupts
>>     312.50 ± 42%     -54.2%     143.00 ± 18%  interrupts.CPU159.RES:Rescheduling_interrupts
>>     725.00 ±118%     -75.7%     176.25 ± 14%  interrupts.CPU163.RES:Rescheduling_interrupts
>>       2367 ±  6%     +59.9%       3786 ± 24%  interrupts.CPU177.NMI:Non-maskable_interrupts
>>       2367 ±  6%     +59.9%       3786 ± 24%  interrupts.CPU177.PMI:Performance_monitoring_interrupts
>>     239.50 ± 30%     -46.6%     128.00 ± 14%  interrupts.CPU179.RES:Rescheduling_interrupts
>>     320.75 ± 15%     -24.0%     243.75 ± 20%  interrupts.CPU20.RES:Rescheduling_interrupts
>>     302.50 ± 17%     -47.2%     159.75 ±  8%  interrupts.CPU200.RES:Rescheduling_interrupts
>>       2166 ±  5%     +92.0%       4157 ± 40%  interrupts.CPU207.NMI:Non-maskable_interrupts
>>       2166 ±  5%     +92.0%       4157 ± 40%  interrupts.CPU207.PMI:Performance_monitoring_interrupts
>>     217.00 ± 11%     -34.6%     142.00 ± 12%  interrupts.CPU214.RES:Rescheduling_interrupts
>>       2610 ± 36%     +47.4%       3848 ± 35%  interrupts.CPU215.NMI:Non-maskable_interrupts
>>       2610 ± 36%     +47.4%       3848 ± 35%  interrupts.CPU215.PMI:Performance_monitoring_interrupts
>>       2046 ± 13%    +118.6%       4475 ± 43%  interrupts.CPU22.NMI:Non-maskable_interrupts
>>       2046 ± 13%    +118.6%       4475 ± 43%  interrupts.CPU22.PMI:Performance_monitoring_interrupts
>>     289.50 ± 28%     -41.1%     170.50 ±  8%  interrupts.CPU22.RES:Rescheduling_interrupts
>>       2232 ±  6%     +33.0%       2970 ± 24%  interrupts.CPU221.NMI:Non-maskable_interrupts
>>       2232 ±  6%     +33.0%       2970 ± 24%  interrupts.CPU221.PMI:Performance_monitoring_interrupts
>>       4552 ± 12%     -27.6%       3295 ± 15%  interrupts.CPU222.NMI:Non-maskable_interrupts
>>       4552 ± 12%     -27.6%       3295 ± 15%  interrupts.CPU222.PMI:Performance_monitoring_interrupts
>>       2013 ± 15%     +80.9%       3641 ± 27%  interrupts.CPU226.NMI:Non-maskable_interrupts
>>       2013 ± 15%     +80.9%       3641 ± 27%  interrupts.CPU226.PMI:Performance_monitoring_interrupts
>>       2575 ± 49%     +67.1%       4302 ± 34%  interrupts.CPU227.NMI:Non-maskable_interrupts
>>       2575 ± 49%     +67.1%       4302 ± 34%  interrupts.CPU227.PMI:Performance_monitoring_interrupts
>>     248.00 ± 36%     -36.3%     158.00 ± 19%  interrupts.CPU228.RES:Rescheduling_interrupts
>>       2441 ± 24%     +43.0%       3490 ± 30%  interrupts.CPU23.NMI:Non-maskable_interrupts
>>       2441 ± 24%     +43.0%       3490 ± 30%  interrupts.CPU23.PMI:Performance_monitoring_interrupts
>>     404.25 ± 69%     -65.5%     139.50 ± 17%  interrupts.CPU236.RES:Rescheduling_interrupts
>>     566.50 ± 40%     -73.6%     149.50 ± 31%  interrupts.CPU237.RES:Rescheduling_interrupts
>>     243.50 ± 26%     -37.1%     153.25 ± 21%  interrupts.CPU248.RES:Rescheduling_interrupts
>>     258.25 ± 12%     -53.5%     120.00 ± 18%  interrupts.CPU249.RES:Rescheduling_interrupts
>>       2888 ± 27%     +49.4%       4313 ± 30%  interrupts.CPU253.NMI:Non-maskable_interrupts
>>       2888 ± 27%     +49.4%       4313 ± 30%  interrupts.CPU253.PMI:Performance_monitoring_interrupts
>>       2468 ± 44%     +67.3%       4131 ± 37%  interrupts.CPU256.NMI:Non-maskable_interrupts
>>       2468 ± 44%     +67.3%       4131 ± 37%  interrupts.CPU256.PMI:Performance_monitoring_interrupts
>>     425.00 ± 59%     -60.3%     168.75 ± 34%  interrupts.CPU258.RES:Rescheduling_interrupts
>>       1859 ± 16%    +106.3%       3834 ± 44%  interrupts.CPU268.NMI:Non-maskable_interrupts
>>       1859 ± 16%    +106.3%       3834 ± 44%  interrupts.CPU268.PMI:Performance_monitoring_interrupts
>>       2684 ± 28%     +61.2%       4326 ± 36%  interrupts.CPU269.NMI:Non-maskable_interrupts
>>       2684 ± 28%     +61.2%       4326 ± 36%  interrupts.CPU269.PMI:Performance_monitoring_interrupts
>>       2171 ±  6%    +108.8%       4533 ± 20%  interrupts.CPU270.NMI:Non-maskable_interrupts
>>       2171 ±  6%    +108.8%       4533 ± 20%  interrupts.CPU270.PMI:Performance_monitoring_interrupts
>>       2262 ± 14%     +61.8%       3659 ± 37%  interrupts.CPU273.NMI:Non-maskable_interrupts
>>       2262 ± 14%     +61.8%       3659 ± 37%  interrupts.CPU273.PMI:Performance_monitoring_interrupts
>>       2203 ± 11%     +50.7%       3320 ± 38%  interrupts.CPU279.NMI:Non-maskable_interrupts
>>       2203 ± 11%     +50.7%       3320 ± 38%  interrupts.CPU279.PMI:Performance_monitoring_interrupts
>>       2433 ± 17%     +52.9%       3721 ± 25%  interrupts.CPU280.NMI:Non-maskable_interrupts
>>       2433 ± 17%     +52.9%       3721 ± 25%  interrupts.CPU280.PMI:Performance_monitoring_interrupts
>>       2778 ± 33%     +63.1%       4531 ± 36%  interrupts.CPU283.NMI:Non-maskable_interrupts
>>       2778 ± 33%     +63.1%       4531 ± 36%  interrupts.CPU283.PMI:Performance_monitoring_interrupts
>>     331.75 ± 32%     -39.8%     199.75 ± 17%  interrupts.CPU29.RES:Rescheduling_interrupts
>>       2178 ± 22%     +53.9%       3353 ± 31%  interrupts.CPU3.NMI:Non-maskable_interrupts
>>       2178 ± 22%     +53.9%       3353 ± 31%  interrupts.CPU3.PMI:Performance_monitoring_interrupts
>>     298.50 ± 30%     -39.7%     180.00 ±  6%  interrupts.CPU34.RES:Rescheduling_interrupts
>>       2490 ±  3%     +58.7%       3953 ± 28%  interrupts.CPU35.NMI:Non-maskable_interrupts
>>       2490 ±  3%     +58.7%       3953 ± 28%  interrupts.CPU35.PMI:Performance_monitoring_interrupts
>>     270.50 ± 24%     -31.1%     186.25 ±  3%  interrupts.CPU36.RES:Rescheduling_interrupts
>>       2493 ±  7%     +57.0%       3915 ± 27%  interrupts.CPU43.NMI:Non-maskable_interrupts
>>       2493 ±  7%     +57.0%       3915 ± 27%  interrupts.CPU43.PMI:Performance_monitoring_interrupts
>>     286.75 ± 36%     -32.4%     193.75 ±  7%  interrupts.CPU45.RES:Rescheduling_interrupts
>>     259.00 ± 12%     -23.6%     197.75 ± 13%  interrupts.CPU46.RES:Rescheduling_interrupts
>>     244.00 ± 21%     -35.6%     157.25 ± 11%  interrupts.CPU47.RES:Rescheduling_interrupts
>>     230.00 ±  7%     -21.3%     181.00 ± 11%  interrupts.CPU48.RES:Rescheduling_interrupts
>>     281.00 ± 13%     -27.4%     204.00 ± 15%  interrupts.CPU53.RES:Rescheduling_interrupts
>>     256.75 ±  5%     -18.4%     209.50 ± 12%  interrupts.CPU54.RES:Rescheduling_interrupts
>>       2433 ±  9%     +68.4%       4098 ± 35%  interrupts.CPU58.NMI:Non-maskable_interrupts
>>       2433 ±  9%     +68.4%       4098 ± 35%  interrupts.CPU58.PMI:Performance_monitoring_interrupts
>>     316.00 ± 25%     -41.4%     185.25 ± 13%  interrupts.CPU59.RES:Rescheduling_interrupts
>>       2703 ± 38%     +56.0%       4217 ± 31%  interrupts.CPU60.NMI:Non-maskable_interrupts
>>       2703 ± 38%     +56.0%       4217 ± 31%  interrupts.CPU60.PMI:Performance_monitoring_interrupts
>>       2425 ± 16%     +39.9%       3394 ± 27%  interrupts.CPU61.NMI:Non-maskable_interrupts
>>       2425 ± 16%     +39.9%       3394 ± 27%  interrupts.CPU61.PMI:Performance_monitoring_interrupts
>>       2388 ± 18%     +69.5%       4047 ± 29%  interrupts.CPU66.NMI:Non-maskable_interrupts
>>       2388 ± 18%     +69.5%       4047 ± 29%  interrupts.CPU66.PMI:Performance_monitoring_interrupts
>>       2322 ± 11%     +93.4%       4491 ± 35%  interrupts.CPU67.NMI:Non-maskable_interrupts
>>       2322 ± 11%     +93.4%       4491 ± 35%  interrupts.CPU67.PMI:Performance_monitoring_interrupts
>>     319.00 ± 40%     -44.7%     176.25 ±  9%  interrupts.CPU67.RES:Rescheduling_interrupts
>>       2512 ±  8%     +28.1%       3219 ± 25%  interrupts.CPU70.NMI:Non-maskable_interrupts
>>       2512 ±  8%     +28.1%       3219 ± 25%  interrupts.CPU70.PMI:Performance_monitoring_interrupts
>>       2290 ± 39%     +78.7%       4094 ± 28%  interrupts.CPU74.NMI:Non-maskable_interrupts
>>       2290 ± 39%     +78.7%       4094 ± 28%  interrupts.CPU74.PMI:Performance_monitoring_interrupts
>>       2446 ± 40%     +94.8%       4764 ± 23%  interrupts.CPU75.NMI:Non-maskable_interrupts
>>       2446 ± 40%     +94.8%       4764 ± 23%  interrupts.CPU75.PMI:Performance_monitoring_interrupts
>>     426.75 ± 61%     -67.7%     138.00 ±  8%  interrupts.CPU75.RES:Rescheduling_interrupts
>>     192.50 ± 13%     +45.6%     280.25 ± 45%  interrupts.CPU76.RES:Rescheduling_interrupts
>>     274.25 ± 34%     -42.2%     158.50 ± 34%  interrupts.CPU77.RES:Rescheduling_interrupts
>>       2357 ±  9%     +73.0%       4078 ± 23%  interrupts.CPU78.NMI:Non-maskable_interrupts
>>       2357 ±  9%     +73.0%       4078 ± 23%  interrupts.CPU78.PMI:Performance_monitoring_interrupts
>>     348.50 ± 53%     -47.3%     183.75 ± 29%  interrupts.CPU80.RES:Rescheduling_interrupts
>>       2650 ± 43%     +46.2%       3874 ± 36%  interrupts.CPU84.NMI:Non-maskable_interrupts
>>       2650 ± 43%     +46.2%       3874 ± 36%  interrupts.CPU84.PMI:Performance_monitoring_interrupts
>>       2235 ± 10%    +117.8%       4867 ± 10%  interrupts.CPU90.NMI:Non-maskable_interrupts
>>       2235 ± 10%    +117.8%       4867 ± 10%  interrupts.CPU90.PMI:Performance_monitoring_interrupts
>>       2606 ± 33%     +38.1%       3598 ± 21%  interrupts.CPU92.NMI:Non-maskable_interrupts
>>       2606 ± 33%     +38.1%       3598 ± 21%  interrupts.CPU92.PMI:Performance_monitoring_interrupts
>>     408.75 ± 58%     -56.8%     176.75 ± 25%  interrupts.CPU92.RES:Rescheduling_interrupts
>>     399.00 ± 64%     -63.6%     145.25 ± 16%  interrupts.CPU93.RES:Rescheduling_interrupts
>>     314.75 ± 36%     -44.2%     175.75 ± 13%  interrupts.CPU94.RES:Rescheduling_interrupts
>>     191.00 ± 15%     -29.1%     135.50 ±  9%  interrupts.CPU97.RES:Rescheduling_interrupts
>>      94.00 ±  8%     +50.0%     141.00 ± 12%  interrupts.IWI:IRQ_work_interrupts
>>     841457 ±  7%     +16.6%     980751 ±  3%  interrupts.NMI:Non-maskable_interrupts
>>     841457 ±  7%     +16.6%     980751 ±  3%  interrupts.PMI:Performance_monitoring_interrupts
>>      12.75 ± 11%      -4.1        8.67 ± 31%  perf-profile.calltrace.cycles-pp.do_rw_once
>>       1.02 ± 16%      -0.6        0.47 ± 59%  perf-profile.calltrace.cycles-pp.sched_clock.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter.do_idle
>>       1.10 ± 15%      -0.4        0.66 ± 14%  perf-profile.calltrace.cycles-pp.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>>       1.05 ± 16%      -0.4        0.61 ± 14%  perf-profile.calltrace.cycles-pp.native_sched_clock.sched_clock.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter
>>       1.58 ±  4%      +0.3        1.91 ±  7%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page
>>       0.79 ± 26%      +0.5        1.27 ± 18%  perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>       0.79 ± 26%      +0.5        1.27 ± 18%  perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>       0.79 ± 26%      +0.5        1.27 ± 18%  perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>       2.11 ±  4%      +0.5        2.60 ±  7%  perf-profile.calltrace.cycles-pp.apic_timer_interrupt.osq_lock.__mutex_lock.hugetlb_fault.handle_mm_fault
>>       0.83 ± 26%      +0.5        1.32 ± 18%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>>       0.83 ± 26%      +0.5        1.32 ± 18%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>>       1.90 ±  5%      +0.6        2.45 ±  7%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page.copy_subpage
>>       0.65 ± 62%      +0.6        1.20 ± 15%  perf-profile.calltrace.cycles-pp.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault
>>       0.60 ± 62%      +0.6        1.16 ± 18%  perf-profile.calltrace.cycles-pp.free_huge_page.release_pages.tlb_flush_mmu.tlb_finish_mmu.exit_mmap
>>       0.95 ± 17%      +0.6        1.52 ±  8%  perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner
>>       0.61 ± 62%      +0.6        1.18 ± 18%  perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.exit_mmap.mmput
>>       0.61 ± 62%      +0.6        1.19 ± 19%  perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.mmput.do_exit.do_group_exit
>>       0.61 ± 62%      +0.6        1.19 ± 19%  perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.exit_mmap.mmput.do_exit
>>       0.64 ± 61%      +0.6        1.23 ± 18%  perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>>       0.64 ± 61%      +0.6        1.23 ± 18%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group
>>       1.30 ±  9%      +0.6        1.92 ±  8%  perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock
>>       0.19 ±173%      +0.7        0.89 ± 20%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_huge_page.release_pages.tlb_flush_mmu
>>       0.19 ±173%      +0.7        0.90 ± 20%  perf-profile.calltrace.cycles-pp._raw_spin_lock.free_huge_page.release_pages.tlb_flush_mmu.tlb_finish_mmu
>>       0.00            +0.8        0.77 ± 30%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page
>>       0.00            +0.8        0.78 ± 30%  perf-profile.calltrace.cycles-pp._raw_spin_lock.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page
>>       0.00            +0.8        0.79 ± 29%  perf-profile.calltrace.cycles-pp.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow
>>       0.82 ± 67%      +0.9        1.72 ± 22%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.alloc_huge_page.hugetlb_cow.hugetlb_fault
>>       0.84 ± 66%      +0.9        1.74 ± 20%  perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow
>>       2.52 ±  6%      +0.9        3.44 ±  9%  perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page.copy_subpage.copy_user_huge_page
>>       0.83 ± 67%      +0.9        1.75 ± 21%  perf-profile.calltrace.cycles-pp._raw_spin_lock.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault
>>       0.84 ± 66%      +0.9        1.77 ± 20%  perf-profile.calltrace.cycles-pp._raw_spin_lock.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault
>>       1.64 ± 12%      +1.0        2.67 ±  7%  perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock.hugetlb_fault
>>       1.65 ± 45%      +1.3        2.99 ± 18%  perf-profile.calltrace.cycles-pp.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault
>>       1.74 ± 13%      +1.4        3.16 ±  6%  perf-profile.calltrace.cycles-pp.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock.hugetlb_fault.handle_mm_fault
>>       2.56 ± 48%      +2.2        4.81 ± 19%  perf-profile.calltrace.cycles-pp.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault.__do_page_fault
>>      12.64 ± 14%      +3.6       16.20 ±  8%  perf-profile.calltrace.cycles-pp.mutex_spin_on_owner.__mutex_lock.hugetlb_fault.handle_mm_fault.__do_page_fault
>>       2.97 ±  7%      +3.8        6.74 ±  9%  perf-profile.calltrace.cycles-pp.apic_timer_interrupt.copy_page.copy_subpage.copy_user_huge_page.hugetlb_cow
>>      19.99 ±  9%      +4.1       24.05 ±  6%  perf-profile.calltrace.cycles-pp.hugetlb_cow.hugetlb_fault.handle_mm_fault.__do_page_fault.do_page_fault
>>       1.37 ± 15%      -0.5        0.83 ± 13%  perf-profile.children.cycles-pp.sched_clock_cpu
>>       1.31 ± 16%      -0.5        0.78 ± 13%  perf-profile.children.cycles-pp.sched_clock
>>       1.29 ± 16%      -0.5        0.77 ± 13%  perf-profile.children.cycles-pp.native_sched_clock
>>       1.80 ±  2%      -0.3        1.47 ± 10%  perf-profile.children.cycles-pp.task_tick_fair
>>       0.73 ±  2%      -0.2        0.54 ± 11%  perf-profile.children.cycles-pp.update_curr
>>       0.42 ± 17%      -0.2        0.27 ± 16%  perf-profile.children.cycles-pp.account_process_tick
>>       0.73 ± 10%      -0.2        0.58 ±  9%  perf-profile.children.cycles-pp.rcu_sched_clock_irq
>>       0.27 ±  6%      -0.1        0.14 ± 14%  perf-profile.children.cycles-pp.__acct_update_integrals
>>       0.27 ± 18%      -0.1        0.16 ± 13%  perf-profile.children.cycles-pp.rcu_segcblist_ready_cbs
>>       0.40 ± 12%      -0.1        0.30 ± 14%  perf-profile.children.cycles-pp.__next_timer_interrupt
>>       0.47 ±  7%      -0.1        0.39 ± 13%  perf-profile.children.cycles-pp.update_rq_clock
>>       0.29 ± 12%      -0.1        0.21 ± 15%  perf-profile.children.cycles-pp.cpuidle_governor_latency_req
>>       0.21 ±  7%      -0.1        0.14 ± 12%  perf-profile.children.cycles-pp.account_system_index_time
>>       0.38 ±  2%      -0.1        0.31 ± 12%  perf-profile.children.cycles-pp.timerqueue_add
>>       0.26 ± 11%      -0.1        0.20 ± 13%  perf-profile.children.cycles-pp.find_next_bit
>>       0.23 ± 15%      -0.1        0.17 ± 15%  perf-profile.children.cycles-pp.rcu_dynticks_eqs_exit
>>       0.14 ±  8%      -0.1        0.07 ± 14%  perf-profile.children.cycles-pp.account_user_time
>>       0.17 ±  6%      -0.0        0.12 ± 10%  perf-profile.children.cycles-pp.cpuacct_charge
>>       0.18 ± 20%      -0.0        0.13 ±  3%  perf-profile.children.cycles-pp.irq_work_tick
>>       0.11 ± 13%      -0.0        0.07 ± 25%  perf-profile.children.cycles-pp.tick_sched_do_timer
>>       0.12 ± 10%      -0.0        0.08 ± 15%  perf-profile.children.cycles-pp.get_cpu_device
>>       0.07 ± 11%      -0.0        0.04 ± 58%  perf-profile.children.cycles-pp.raise_softirq
>>       0.12 ±  3%      -0.0        0.09 ±  8%  perf-profile.children.cycles-pp.write
>>       0.11 ± 13%      +0.0        0.14 ±  8%  perf-profile.children.cycles-pp.native_write_msr
>>       0.09 ±  9%      +0.0        0.11 ±  7%  perf-profile.children.cycles-pp.finish_task_switch
>>       0.10 ± 10%      +0.0        0.13 ±  5%  perf-profile.children.cycles-pp.schedule_idle
>>       0.07 ±  6%      +0.0        0.10 ± 12%  perf-profile.children.cycles-pp.__read_nocancel
>>       0.04 ± 58%      +0.0        0.07 ± 15%  perf-profile.children.cycles-pp.__free_pages_ok
>>       0.06 ±  7%      +0.0        0.09 ± 13%  perf-profile.children.cycles-pp.perf_read
>>       0.07            +0.0        0.11 ± 14%  perf-profile.children.cycles-pp.perf_evsel__read_counter
>>       0.07            +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.cmd_stat
>>       0.07            +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.__run_perf_stat
>>       0.07            +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.process_interval
>>       0.07            +0.0        0.11 ± 13%  perf-profile.children.cycles-pp.read_counters
>>       0.07 ± 22%      +0.0        0.11 ± 19%  perf-profile.children.cycles-pp.__handle_mm_fault
>>       0.07 ± 19%      +0.1        0.13 ±  8%  perf-profile.children.cycles-pp.rb_erase
>>       0.03 ±100%      +0.1        0.09 ±  9%  perf-profile.children.cycles-pp.smp_call_function_single
>>       0.01 ±173%      +0.1        0.08 ± 11%  perf-profile.children.cycles-pp.perf_event_read
>>       0.00            +0.1        0.07 ± 13%  perf-profile.children.cycles-pp.__perf_event_read_value
>>       0.00            +0.1        0.07 ±  7%  perf-profile.children.cycles-pp.__intel_pmu_enable_all
>>       0.08 ± 17%      +0.1        0.15 ±  8%  perf-profile.children.cycles-pp.native_apic_msr_eoi_write
>>       0.04 ±103%      +0.1        0.13 ± 58%  perf-profile.children.cycles-pp.shmem_getpage_gfp
>>       0.38 ± 14%      +0.1        0.51 ±  6%  perf-profile.children.cycles-pp.run_timer_softirq
>>       0.11 ±  4%      +0.3        0.37 ± 32%  perf-profile.children.cycles-pp.worker_thread
>>       0.20 ±  5%      +0.3        0.48 ± 25%  perf-profile.children.cycles-pp.ret_from_fork
>>       0.20 ±  4%      +0.3        0.48 ± 25%  perf-profile.children.cycles-pp.kthread
>>       0.00            +0.3        0.29 ± 38%  perf-profile.children.cycles-pp.memcpy_erms
>>       0.00            +0.3        0.29 ± 38%  perf-profile.children.cycles-pp.drm_fb_helper_dirty_work
>>       0.00            +0.3        0.31 ± 37%  perf-profile.children.cycles-pp.process_one_work
>>       0.47 ± 48%      +0.4        0.91 ± 19%  perf-profile.children.cycles-pp.prep_new_huge_page
>>       0.70 ± 29%      +0.5        1.16 ± 18%  perf-profile.children.cycles-pp.free_huge_page
>>       0.73 ± 29%      +0.5        1.19 ± 18%  perf-profile.children.cycles-pp.tlb_flush_mmu
>>       0.72 ± 29%      +0.5        1.18 ± 18%  perf-profile.children.cycles-pp.release_pages
>>       0.73 ± 29%      +0.5        1.19 ± 18%  perf-profile.children.cycles-pp.tlb_finish_mmu
>>       0.76 ± 27%      +0.5        1.23 ± 18%  perf-profile.children.cycles-pp.exit_mmap
>>       0.77 ± 27%      +0.5        1.24 ± 18%  perf-profile.children.cycles-pp.mmput
>>       0.79 ± 26%      +0.5        1.27 ± 18%  perf-profile.children.cycles-pp.__x64_sys_exit_group
>>       0.79 ± 26%      +0.5        1.27 ± 18%  perf-profile.children.cycles-pp.do_group_exit
>>       0.79 ± 26%      +0.5        1.27 ± 18%  perf-profile.children.cycles-pp.do_exit
>>       1.28 ± 29%      +0.5        1.76 ±  9%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
>>       0.77 ± 28%      +0.5        1.26 ± 13%  perf-profile.children.cycles-pp.alloc_fresh_huge_page
>>       1.53 ± 15%      +0.7        2.26 ± 14%  perf-profile.children.cycles-pp.do_syscall_64
>>       1.53 ± 15%      +0.7        2.27 ± 14%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>>       1.13 ±  3%      +0.9        2.07 ± 14%  perf-profile.children.cycles-pp.interrupt_entry
>>       0.79 ±  9%      +1.0        1.76 ±  5%  perf-profile.children.cycles-pp.perf_event_task_tick
>>       1.71 ± 39%      +1.4        3.08 ± 16%  perf-profile.children.cycles-pp.alloc_surplus_huge_page
>>       2.66 ± 42%      +2.3        4.94 ± 17%  perf-profile.children.cycles-pp.alloc_huge_page
>>       2.89 ± 45%      +2.7        5.54 ± 18%  perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>>       3.34 ± 35%      +2.7        6.02 ± 17%  perf-profile.children.cycles-pp._raw_spin_lock
>>      12.77 ± 14%      +3.9       16.63 ±  7%  perf-profile.children.cycles-pp.mutex_spin_on_owner
>>      20.12 ±  9%      +4.0       24.16 ±  6%  perf-profile.children.cycles-pp.hugetlb_cow
>>      15.40 ± 10%      -3.6       11.84 ± 28%  perf-profile.self.cycles-pp.do_rw_once
>>       4.02 ±  9%      -1.3        2.73 ± 30%  perf-profile.self.cycles-pp.do_access
>>       2.00 ± 14%      -0.6        1.41 ± 13%  perf-profile.self.cycles-pp.cpuidle_enter_state
>>       1.26 ± 16%      -0.5        0.74 ± 13%  perf-profile.self.cycles-pp.native_sched_clock
>>       0.42 ± 17%      -0.2        0.27 ± 16%  perf-profile.self.cycles-pp.account_process_tick
>>       0.27 ± 19%      -0.2        0.12 ± 17%  perf-profile.self.cycles-pp.timerqueue_del
>>       0.53 ±  3%      -0.1        0.38 ± 11%  perf-profile.self.cycles-pp.update_curr
>>       0.27 ±  6%      -0.1        0.14 ± 14%  perf-profile.self.cycles-pp.__acct_update_integrals
>>       0.27 ± 18%      -0.1        0.16 ± 13%  perf-profile.self.cycles-pp.rcu_segcblist_ready_cbs
>>       0.61 ±  4%      -0.1        0.51 ±  8%  perf-profile.self.cycles-pp.task_tick_fair
>>       0.20 ±  8%      -0.1        0.12 ± 14%  perf-profile.self.cycles-pp.account_system_index_time
>>       0.23 ± 15%      -0.1        0.16 ± 17%  perf-profile.self.cycles-pp.rcu_dynticks_eqs_exit
>>       0.25 ± 11%      -0.1        0.18 ± 14%  perf-profile.self.cycles-pp.find_next_bit
>>       0.10 ± 11%      -0.1        0.03 ±100%  perf-profile.self.cycles-pp.tick_sched_do_timer
>>       0.29            -0.1        0.23 ± 11%  perf-profile.self.cycles-pp.timerqueue_add
>>       0.12 ± 10%      -0.1        0.06 ± 17%  perf-profile.self.cycles-pp.account_user_time
>>       0.22 ± 15%      -0.1        0.16 ±  6%  perf-profile.self.cycles-pp.scheduler_tick
>>       0.17 ±  6%      -0.0        0.12 ± 10%  perf-profile.self.cycles-pp.cpuacct_charge
>>       0.18 ± 20%      -0.0        0.13 ±  3%  perf-profile.self.cycles-pp.irq_work_tick
>>       0.07 ± 13%      -0.0        0.03 ±100%  perf-profile.self.cycles-pp.update_process_times
>>       0.12 ±  7%      -0.0        0.08 ± 15%  perf-profile.self.cycles-pp.get_cpu_device
>>       0.07 ± 11%      -0.0        0.04 ± 58%  perf-profile.self.cycles-pp.raise_softirq
>>       0.12 ± 11%      -0.0        0.09 ±  7%  perf-profile.self.cycles-pp.tick_nohz_get_sleep_length
>>       0.11 ± 11%      +0.0        0.14 ±  6%  perf-profile.self.cycles-pp.native_write_msr
>>       0.10 ±  5%      +0.1        0.15 ±  8%  perf-profile.self.cycles-pp.__remove_hrtimer
>>       0.07 ± 23%      +0.1        0.13 ±  8%  perf-profile.self.cycles-pp.rb_erase
>>       0.08 ± 17%      +0.1        0.15 ±  7%  perf-profile.self.cycles-pp.native_apic_msr_eoi_write
>>       0.00            +0.1        0.08 ± 10%  perf-profile.self.cycles-pp.smp_call_function_single
>>       0.32 ± 17%      +0.1        0.42 ±  7%  perf-profile.self.cycles-pp.run_timer_softirq
>>       0.22 ±  5%      +0.1        0.34 ±  4%  perf-profile.self.cycles-pp.ktime_get_update_offsets_now
>>       0.45 ± 15%      +0.2        0.60 ± 12%  perf-profile.self.cycles-pp.rcu_irq_enter
>>       0.31 ±  8%      +0.2        0.46 ± 16%  perf-profile.self.cycles-pp.irq_enter
>>       0.29 ± 10%      +0.2        0.44 ± 16%  perf-profile.self.cycles-pp.apic_timer_interrupt
>>       0.71 ± 30%      +0.2        0.92 ±  8%  perf-profile.self.cycles-pp.perf_mux_hrtimer_handler
>>       0.00            +0.3        0.28 ± 37%  perf-profile.self.cycles-pp.memcpy_erms
>>       1.12 ±  3%      +0.9        2.02 ± 15%  perf-profile.self.cycles-pp.interrupt_entry
>>       0.79 ±  9%      +0.9        1.73 ±  5%  perf-profile.self.cycles-pp.perf_event_task_tick
>>       2.49 ± 45%      +2.1        4.55 ± 20%  perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>>      10.95 ± 15%      +2.7       13.61 ±  8%  perf-profile.self.cycles-pp.mutex_spin_on_owner
>>
>>
>>                                                                                 
>>                                vm-scalability.throughput                        
>>                                                                                 
>>   1.6e+07 +-+---------------------------------------------------------------+   
>>           |..+.+    +..+.+..+.+.   +.      +..+.+..+.+..+.+..+.+..+    +    |   
>>   1.4e+07 +-+  :    :  O      O    O                           O            |   
>>   1.2e+07 O-+O O  O O    O  O    O    O O  O  O    O    O    O      O  O O  O   
>>           |     :   :                           O    O    O       O         |   
>>     1e+07 +-+   :  :                                                        |   
>>           |     :  :                                                        |   
>>     8e+06 +-+   :  :                                                        |   
>>           |      : :                                                        |   
>>     6e+06 +-+    : :                                                        |   
>>     4e+06 +-+    : :                                                        |   
>>           |      ::                                                         |   
>>     2e+06 +-+     :                                                         |   
>>           |       :                                                         |   
>>         0 +-+---------------------------------------------------------------+   
>>                                                                                 
>>                                                                                                                                                                 
>>                          vm-scalability.time.minor_page_faults                  
>>                                                                                 
>>   2.5e+06 +-+---------------------------------------------------------------+   
>>           |                                                                 |   
>>           |..+.+    +..+.+..+.+..+.+..+.+..  .+.  .+.+..+.+..+.+..+.+..+    |   
>>     2e+06 +-+  :    :                      +.   +.                          |   
>>           O  O O: O O  O O  O O  O O                    O      O            |   
>>           |     :   :                 O O  O  O O  O O    O  O    O O  O O  O   
>>   1.5e+06 +-+   :  :                                                        |   
>>           |     :  :                                                        |   
>>     1e+06 +-+    : :                                                        |   
>>           |      : :                                                        |   
>>           |      : :                                                        |   
>>    500000 +-+    : :                                                        |   
>>           |       :                                                         |   
>>           |       :                                                         |   
>>         0 +-+---------------------------------------------------------------+   
>>                                                                                 
>>                                                                                                                                                                 
>>                                 vm-scalability.workload                         
>>                                                                                 
>>   3.5e+09 +-+---------------------------------------------------------------+   
>>           | .+.                      .+.+..                        .+..     |   
>>     3e+09 +-+  +    +..+.+..+.+..+.+.      +..+.+..+.+..+.+..+.+..+    +    |   
>>           |    :    :       O O                                O            |   
>>   2.5e+09 O-+O O: O O  O O       O O  O    O            O                   |   
>>           |     :   :                   O     O O  O O    O  O    O O  O O  O   
>>     2e+09 +-+   :  :                                                        |   
>>           |     :  :                                                        |   
>>   1.5e+09 +-+    : :                                                        |   
>>           |      : :                                                        |   
>>     1e+09 +-+    : :                                                        |   
>>           |      : :                                                        |   
>>     5e+08 +-+     :                                                         |   
>>           |       :                                                         |   
>>         0 +-+---------------------------------------------------------------+   
>>                                                                                 
>>                                                                                 
>> [*] bisect-good sample
>> [O] bisect-bad  sample
>>
>>
>>
>> Disclaimer:
>> Results have been estimated based on internal Intel analysis and are provided
>> for informational purposes only. Any difference in system hardware or software
>> design or configuration may affect actual performance.
>>
>>
>> Thanks,
>> Rong Chen
>>
> 

-- 
Thomas Zimmermann
Graphics Driver Developer
SUSE Linux GmbH, Maxfeldstrasse 5, 90409 Nuernberg, Germany
GF: Felix Imendörffer, Mary Higgins, Sri Rasiah
HRB 21284 (AG Nürnberg)

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 488 bytes
Desc: OpenPGP digital signature
URL: <https://lists.freedesktop.org/archives/dri-devel/attachments/20190804/a3308ceb/attachment-0001.sig>


More information about the dri-devel mailing list