Greeting, FYI, we noticed a 4671.0% improvement of stress-ng.numa.ops_per_sec due to commit: commit: cdaa813278ddc616ee201eacda77f63996b5dd2d ("[PATCH 4/6] mm/gup: track gup-pinned pages") url: https://github.com/0day-ci/linux/commits/john-hubbard-gmail-com/RFC-v2-mm-gup-dma-tracking/20190205-001101 in testcase: stress-ng on test machine: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory with following parameters: nr_threads: 50% disk: 1HDD testtime: 5s class: memory cpufreq_governor: performance ucode: 0xb00002e In addition to that, the commit also has significant impact on the following tests: +------------------+---------------------------------------------------------------------------+ | testcase: change | will-it-scale: ltp.cve-2017-18075.pass -100.0% undefined | | test machine | qemu-system-x86_64 -enable-kvm -cpu SandyBridge -smp 2 -m 8G | | test parameters | cpufreq_governor=performance | | | mode=process | | | nr_task=100% | | | test=futex1 | +------------------+---------------------------------------------------------------------------+ | testcase: change | stress-ng: | | test machine | 192 threads Skylake-4S with 704G memory | | test parameters | class=cpu | | | cpufreq_governor=performance | | | disk=1HDD | | | nr_threads=100% | | | testtime=1s | +------------------+---------------------------------------------------------------------------+ | testcase: change | stress-ng: stress-ng.numa.ops_per_sec 401.7% improvement | | test machine | 272 threads Intel(R) Xeon Phi(TM) CPU 7255 @ 1.10GHz with 112G memory | | test parameters | class=pipe | | | cpufreq_governor=performance | | | disk=1HDD | | | nr_threads=100% | | | testtime=1s | +------------------+---------------------------------------------------------------------------+ | testcase: change | will-it-scale: stress-ng.vm-splice.ops_per_sec -18.5% regression | | test machine | 272 threads Intel(R) Xeon Phi(TM) CPU 7255 @ 1.10GHz with 112G memory | | test parameters | cpufreq_governor=performance | | | mode=process | | | nr_task=100% | | | test=futex2 | | | ucode=0xb00002e | +------------------+---------------------------------------------------------------------------+ | testcase: change | stress-ng: | | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory | | test parameters | class=os | | | cpufreq_governor=performance | | | disk=1HDD | | | nr_threads=100% | | | testtime=1s | +------------------+---------------------------------------------------------------------------+ | testcase: change | stress-ng: stress-ng.futex.ops_per_sec -58.0% regression | | test machine | 272 threads Intel(R) Xeon Phi(TM) CPU 7255 @ 1.10GHz with 112G memory | | test parameters | class=vm | | | cpufreq_governor=performance | | | disk=1HDD | | | nr_threads=100% | | | testtime=1s | | | ucode=0xb00002e | +------------------+---------------------------------------------------------------------------+ | testcase: change | stress-ng: stress-ng.vm-splice.ops -58.3% undefined | | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory | | test parameters | class=pipe | | | cpufreq_governor=performance | | | disk=1HDD | | | nr_threads=100% | | | testtime=60s | | | ucode=0xb00002e | +------------------+---------------------------------------------------------------------------+ | testcase: change | stress-ng: kernel_selftests.memfd.run_fuse_test.sh.pass -100.0% undefined | | test machine | qemu-system-x86_64 -enable-kvm -cpu SandyBridge -smp 2 -m 4G | | test parameters | class=pipe | | | cpufreq_governor=performance | | | disk=1HDD | | | nr_threads=100% | | | testtime=60s | +------------------+---------------------------------------------------------------------------+ | testcase: change | kvm-unit-tests: stress-ng.vm-splice.ops -99.3% undefined | | test machine | 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory | +------------------+---------------------------------------------------------------------------+ Details are as below: --------------------------------------------------------------------------------------------------> To reproduce: git clone https://github.com/intel/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml ========================================================================================= class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode: memory/gcc-7/performance/1HDD/x86_64-rhel-7.2/50%/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3/stress-ng/5s/0xb00002e commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- %stddev %change %stddev \ | \ 17845 ± 2% +25.1% 22326 ± 22% stress-ng.memcpy.ops 3568 ± 2% +25.1% 4464 ± 22% stress-ng.memcpy.ops_per_sec 55.50 ± 2% +3969.8% 2258 ± 3% stress-ng.numa.ops 9.41 ± 4% +4671.0% 449.07 ± 3% stress-ng.numa.ops_per_sec 277857 ± 5% +50.9% 419386 ± 4% stress-ng.time.involuntary_context_switches 326.50 ± 10% -16.2% 273.50 ± 3% stress-ng.vm-addr.ops 65.30 ± 10% -16.2% 54.71 ± 3% stress-ng.vm-addr.ops_per_sec 0.01 ±113% +0.0 0.02 ± 36% mpstat.cpu.iowait% 181260 +15.7% 209701 vmstat.system.in 64963 ± 32% +315.5% 269897 ± 38% numa-numastat.node0.other_node 66083 ± 33% +904.9% 664072 ± 18% numa-numastat.node1.other_node 1670 ± 2% +36.4% 2279 ± 3% slabinfo.numa_policy.active_objs 1670 ± 2% +36.4% 2279 ± 3% slabinfo.numa_policy.num_objs 38695830 +29.5% 50110710 turbostat.IRQ 10.08 -2.9% 9.79 turbostat.RAMWatt 917912 ± 10% +77.1% 1625317 ± 27% meminfo.Active 917736 ± 10% +77.1% 1624942 ± 27% meminfo.Active(anon) 425637 ± 10% +43.8% 611934 ± 18% meminfo.Inactive 425296 ± 10% +43.8% 611561 ± 18% meminfo.Inactive(anon) 3195353 ± 3% +26.7% 4047153 ± 14% meminfo.Memused 49.89 ± 5% -26.7% 36.55 ± 6% perf-stat.i.MPKI 4.372e+08 ± 2% -5.1% 4.151e+08 ± 3% perf-stat.i.cache-references 9.75 ± 5% -8.8% 8.90 ± 3% perf-stat.i.cpi 4.84 ± 3% -8.2% 4.44 ± 4% perf-stat.overall.MPKI 4.361e+08 ± 2% -5.2% 4.137e+08 ± 3% perf-stat.ps.cache-references 27667 ± 19% +32.4% 36627 ± 17% softirqs.CPU0.RCU 24079 ± 6% -12.6% 21054 ± 8% softirqs.CPU24.SCHED 29249 ± 4% +14.0% 33351 ± 4% softirqs.CPU25.RCU 29618 ± 4% +16.8% 34600 ± 8% softirqs.CPU27.RCU 37158 ± 3% -12.6% 32476 ± 12% softirqs.CPU69.RCU 446518 ± 23% +87.1% 835393 ± 18% numa-meminfo.node0.Active 446388 ± 23% +87.1% 835224 ± 18% numa-meminfo.node0.Active(anon) 279797 ± 60% +81.1% 506718 ± 7% numa-meminfo.node0.Inactive 279545 ± 60% +81.2% 506611 ± 7% numa-meminfo.node0.Inactive(anon) 57425 ± 7% +16.8% 67077 ± 3% numa-meminfo.node0.KReclaimable 1666995 ± 17% +37.9% 2299473 ± 4% numa-meminfo.node0.MemUsed 57425 ± 7% +16.8% 67077 ± 3% numa-meminfo.node0.SReclaimable 59307 ± 7% -18.7% 48201 ± 4% numa-meminfo.node1.KReclaimable 59307 ± 7% -18.7% 48201 ± 4% numa-meminfo.node1.SReclaimable 114842 ± 19% +86.4% 214054 ± 15% numa-vmstat.node0.nr_active_anon 71405 ± 61% +78.6% 127550 ± 8% numa-vmstat.node0.nr_inactive_anon 257.00 ± 12% -43.0% 146.50 ± 30% numa-vmstat.node0.nr_isolated_anon 14332 ± 7% +17.2% 16792 ± 3% numa-vmstat.node0.nr_slab_reclaimable 114840 ± 19% +86.4% 214050 ± 15% numa-vmstat.node0.nr_zone_active_anon 71404 ± 61% +78.6% 127549 ± 8% numa-vmstat.node0.nr_zone_inactive_anon 39507 ± 32% +250.0% 138280 ± 37% numa-vmstat.node0.numa_other 313.75 ± 5% -57.1% 134.50 ± 25% numa-vmstat.node1.nr_isolated_anon 14816 ± 7% -18.5% 12069 ± 4% numa-vmstat.node1.nr_slab_reclaimable 172051 ± 7% +177.8% 477913 ± 13% numa-vmstat.node1.numa_other 231722 ± 8% +72.6% 399865 ± 27% proc-vmstat.nr_active_anon 291.00 ± 16% +18.6% 345.00 ± 19% proc-vmstat.nr_anon_transparent_hugepages 107048 ± 8% +43.8% 153932 ± 19% proc-vmstat.nr_inactive_anon 643.25 ± 6% -55.9% 283.75 ± 25% proc-vmstat.nr_isolated_anon 231722 ± 8% +72.6% 399865 ± 27% proc-vmstat.nr_zone_active_anon 107048 ± 8% +43.8% 153932 ± 19% proc-vmstat.nr_zone_inactive_anon 131052 ± 3% +612.7% 933974 ± 3% proc-vmstat.numa_other 6.265e+08 ± 3% -13.6% 5.411e+08 ± 5% proc-vmstat.pgalloc_normal 6.264e+08 ± 3% -13.7% 5.403e+08 ± 5% proc-vmstat.pgfree 7486 ±122% +2087.9% 163793 ± 6% proc-vmstat.pgmigrate_fail 1047865 ± 18% +134.0% 2452064 ± 15% proc-vmstat.pgmigrate_success 234795 ± 26% -60.5% 92683 ± 78% proc-vmstat.thp_deferred_split_page 417.00 ±145% +3214.7% 13822 ±144% proc-vmstat.unevictable_pgs_cleared 417.50 ±144% +3211.2% 13824 ±144% proc-vmstat.unevictable_pgs_stranded 46.55 -46.6 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_move_pages.do_syscall_64.entry_SYSCALL_64_after_hwframe 46.55 -46.6 0.00 perf-profile.calltrace.cycles-pp.kernel_move_pages.__x64_sys_move_pages.do_syscall_64.entry_SYSCALL_64_after_hwframe 46.47 -46.5 0.00 perf-profile.calltrace.cycles-pp.do_move_pages_to_node.kernel_move_pages.__x64_sys_move_pages.do_syscall_64.entry_SYSCALL_64_after_hwframe 46.47 -46.5 0.00 perf-profile.calltrace.cycles-pp.migrate_pages.do_move_pages_to_node.kernel_move_pages.__x64_sys_move_pages.do_syscall_64 45.32 -45.3 0.00 perf-profile.calltrace.cycles-pp.move_to_new_page.migrate_pages.do_move_pages_to_node.kernel_move_pages.__x64_sys_move_pages 45.32 -45.3 0.00 perf-profile.calltrace.cycles-pp.migrate_page.move_to_new_page.migrate_pages.do_move_pages_to_node.kernel_move_pages 45.22 -45.2 0.00 perf-profile.calltrace.cycles-pp.migrate_page_copy.migrate_page.move_to_new_page.migrate_pages.do_move_pages_to_node 43.53 -43.0 0.55 ± 62% perf-profile.calltrace.cycles-pp.copy_page.migrate_page_copy.migrate_page.move_to_new_page.migrate_pages 78.97 ± 2% -6.3 72.65 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe 79.20 ± 2% -6.2 72.97 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe 5.01 ± 3% -0.9 4.11 ± 3% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe 5.01 ± 3% -0.9 4.12 ± 3% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe 4.94 ± 3% -0.9 4.05 ± 3% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64 4.99 ± 3% -0.9 4.10 ± 3% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe 2.48 ± 3% -0.5 2.02 ± 3% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap 2.47 ± 3% -0.5 2.02 ± 3% perf-profile.calltrace.cycles-pp.arch_tlb_finish_mmu.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap 2.37 ± 3% -0.4 1.93 ± 3% perf-profile.calltrace.cycles-pp.tlb_flush_mmu_free.arch_tlb_finish_mmu.tlb_finish_mmu.unmap_region.__do_munmap 2.35 ± 3% -0.4 1.92 ± 3% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu_free.arch_tlb_finish_mmu.tlb_finish_mmu.unmap_region 2.38 ± 2% -0.4 1.96 ± 3% perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap 2.38 ± 3% -0.4 1.96 ± 3% perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap.__vm_munmap 2.38 ± 3% -0.4 1.96 ± 3% perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap 2.21 ± 3% -0.4 1.80 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.release_pages.tlb_flush_mmu_free.arch_tlb_finish_mmu.tlb_finish_mmu 2.19 ± 3% -0.4 1.78 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.release_pages.tlb_flush_mmu_free.arch_tlb_finish_mmu 2.23 ± 2% -0.4 1.83 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.pagevec_lru_move_fn.lru_add_drain_cpu.lru_add_drain.unmap_region 2.20 ± 2% -0.4 1.81 ± 3% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.pagevec_lru_move_fn.lru_add_drain_cpu.lru_add_drain 0.58 ± 5% +0.2 0.77 ± 4% perf-profile.calltrace.cycles-pp.touch_atime.pipe_read.__vfs_read.vfs_read.ksys_read 0.58 ± 7% +0.2 0.78 ± 4% perf-profile.calltrace.cycles-pp.anon_pipe_buf_release.pipe_read.__vfs_read.vfs_read.ksys_read 0.70 ± 16% +0.2 0.92 ± 6% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.pipe_read 0.65 ± 5% +0.2 0.88 perf-profile.calltrace.cycles-pp.selinux_file_permission.security_file_permission.vfs_write.ksys_write.do_syscall_64 0.66 ± 5% +0.2 0.89 ± 2% perf-profile.calltrace.cycles-pp.selinux_file_permission.security_file_permission.vfs_read.ksys_read.do_syscall_64 0.70 ± 16% +0.2 0.94 ± 6% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.pipe_read.__vfs_read 0.81 ± 5% +0.3 1.10 ± 3% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.__wake_up_common_lock.pipe_read.__vfs_read.vfs_read 0.83 ± 14% +0.3 1.12 ± 4% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.pipe_read.__vfs_read.vfs_read 0.40 ± 57% +0.3 0.71 perf-profile.calltrace.cycles-pp.__wake_up_common_lock.pipe_write.__vfs_write.vfs_write.ksys_write 0.58 ± 5% +0.3 0.92 ± 28% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin.copy_page_from_iter.pipe_write.__vfs_write 0.28 ±100% +0.4 0.64 ± 8% perf-profile.calltrace.cycles-pp.pipe_wait.pipe_write.__vfs_write.vfs_write.ksys_write 0.65 ± 5% +0.4 1.02 ± 25% perf-profile.calltrace.cycles-pp.copyin.copy_page_from_iter.pipe_write.__vfs_write.vfs_write 0.26 ±100% +0.4 0.65 ± 4% perf-profile.calltrace.cycles-pp.atime_needs_update.touch_atime.pipe_read.__vfs_read.vfs_read 1.04 ± 5% +0.4 1.44 perf-profile.calltrace.cycles-pp.mutex_lock.pipe_write.__vfs_write.vfs_write.ksys_write 1.06 ± 5% +0.4 1.46 ± 2% perf-profile.calltrace.cycles-pp.mutex_lock.pipe_read.__vfs_read.vfs_read.ksys_read 1.21 ± 5% +0.4 1.64 perf-profile.calltrace.cycles-pp.security_file_permission.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.14 ± 8% +0.5 1.61 perf-profile.calltrace.cycles-pp.mutex_unlock.pipe_read.__vfs_read.vfs_read.ksys_read 1.30 ± 5% +0.5 1.78 ± 3% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout.copy_page_to_iter.pipe_read.__vfs_read 1.16 ± 9% +0.5 1.64 perf-profile.calltrace.cycles-pp.mutex_spin_on_owner.__mutex_lock.pipe_write.__vfs_write.vfs_write 0.13 ±173% +0.5 0.63 ± 7% perf-profile.calltrace.cycles-pp.poll_idle.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary 1.37 ± 6% +0.5 1.86 ± 2% perf-profile.calltrace.cycles-pp.copyout.copy_page_to_iter.pipe_read.__vfs_read.vfs_read 1.34 ± 6% +0.5 1.85 perf-profile.calltrace.cycles-pp.security_file_permission.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.13 ±173% +0.5 0.64 perf-profile.calltrace.cycles-pp.file_has_perm.security_file_permission.vfs_write.ksys_write.do_syscall_64 0.14 ±173% +0.5 0.67 ± 2% perf-profile.calltrace.cycles-pp.file_has_perm.security_file_permission.vfs_read.ksys_read.do_syscall_64 0.00 +0.5 0.54 ± 5% perf-profile.calltrace.cycles-pp.rmap_walk_file.remove_migration_ptes.migrate_pages.migrate_to_node.do_migrate_pages 1.25 ± 4% +0.6 1.82 ± 14% perf-profile.calltrace.cycles-pp.copy_page_from_iter.pipe_write.__vfs_write.vfs_write.ksys_write 0.00 +0.6 0.59 ± 4% perf-profile.calltrace.cycles-pp.fsnotify.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +0.6 0.59 ± 6% perf-profile.calltrace.cycles-pp.rmap_walk_anon.try_to_unmap.migrate_pages.migrate_to_node.do_migrate_pages 0.00 +0.6 0.60 ± 8% perf-profile.calltrace.cycles-pp.find_next_bit.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64 1.65 ± 7% +0.6 2.27 ± 4% perf-profile.calltrace.cycles-pp.mutex_unlock.pipe_write.__vfs_write.vfs_write.ksys_write 0.00 +0.6 0.63 ± 6% perf-profile.calltrace.cycles-pp._raw_spin_lock.queue_pages_pte_range.__walk_page_range.walk_page_range.queue_pages_range 0.00 +0.7 0.65 ± 5% perf-profile.calltrace.cycles-pp.remove_migration_ptes.migrate_pages.migrate_to_node.do_migrate_pages.kernel_migrate_pages 1.89 ± 7% +0.7 2.56 perf-profile.calltrace.cycles-pp.__wake_up_common_lock.pipe_read.__vfs_read.vfs_read.ksys_read 0.00 +0.7 0.67 ± 5% perf-profile.calltrace.cycles-pp.queue_pages_test_walk.walk_page_range.queue_pages_range.migrate_to_node.do_migrate_pages 2.05 ± 5% +0.7 2.80 ± 2% perf-profile.calltrace.cycles-pp.copy_page_to_iter.pipe_read.__vfs_read.vfs_read.ksys_read 0.00 +0.8 0.82 ± 23% perf-profile.calltrace.cycles-pp.migrate_page.move_to_new_page.migrate_pages.migrate_to_node.do_migrate_pages 0.00 +0.8 0.82 ± 23% perf-profile.calltrace.cycles-pp.move_to_new_page.migrate_pages.migrate_to_node.do_migrate_pages.kernel_migrate_pages 0.00 +0.8 0.85 ± 5% perf-profile.calltrace.cycles-pp._vm_normal_page.queue_pages_pte_range.__walk_page_range.walk_page_range.queue_pages_range 2.79 ± 4% +0.9 3.68 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64 0.00 +0.9 0.89 ± 3% perf-profile.calltrace.cycles-pp.bitmap_ord_to_pos.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64 2.80 ± 6% +0.9 3.73 perf-profile.calltrace.cycles-pp.syscall_return_via_sysret 0.00 +1.2 1.21 ± 7% perf-profile.calltrace.cycles-pp.smp_call_function_single.on_each_cpu_mask.on_each_cpu_cond_mask.flush_tlb_mm_range.ptep_clear_flush 0.00 +1.3 1.25 ± 7% perf-profile.calltrace.cycles-pp.on_each_cpu_mask.on_each_cpu_cond_mask.flush_tlb_mm_range.ptep_clear_flush.try_to_unmap_one 0.00 +1.4 1.38 ± 7% perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.flush_tlb_mm_range.ptep_clear_flush.try_to_unmap_one.rmap_walk_file 3.73 ± 7% +1.4 5.11 ± 2% perf-profile.calltrace.cycles-pp.__mutex_lock.pipe_write.__vfs_write.vfs_write.ksys_write 0.00 +1.6 1.56 ± 7% perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.ptep_clear_flush.try_to_unmap_one.rmap_walk_file.try_to_unmap 0.00 +1.6 1.63 ± 7% perf-profile.calltrace.cycles-pp.ptep_clear_flush.try_to_unmap_one.rmap_walk_file.try_to_unmap.migrate_pages 0.00 +2.2 2.20 ± 6% perf-profile.calltrace.cycles-pp.try_to_unmap_one.rmap_walk_file.try_to_unmap.migrate_pages.migrate_to_node 0.00 +2.3 2.30 ± 6% perf-profile.calltrace.cycles-pp.rmap_walk_file.try_to_unmap.migrate_pages.migrate_to_node.do_migrate_pages 0.00 +2.9 2.90 ± 6% perf-profile.calltrace.cycles-pp.try_to_unmap.migrate_pages.migrate_to_node.do_migrate_pages.kernel_migrate_pages 8.94 ± 5% +3.4 12.33 perf-profile.calltrace.cycles-pp.pipe_read.__vfs_read.vfs_read.ksys_read.do_syscall_64 9.24 ± 5% +3.5 12.73 perf-profile.calltrace.cycles-pp.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 13.65 ± 18% +3.7 17.34 ± 6% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64 13.65 ± 18% +3.7 17.35 ± 6% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64 13.65 ± 18% +3.7 17.35 ± 6% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64 13.89 ± 17% +3.9 17.84 ± 6% perf-profile.calltrace.cycles-pp.secondary_startup_64 10.58 ± 5% +4.0 14.54 perf-profile.calltrace.cycles-pp.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64 10.90 ± 5% +4.1 14.95 perf-profile.calltrace.cycles-pp.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 11.36 ± 5% +4.3 15.62 perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 11.83 ± 5% +4.4 16.24 perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +4.5 4.54 ± 5% perf-profile.calltrace.cycles-pp.migrate_pages.migrate_to_node.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages 12.69 ± 5% +4.7 17.35 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 13.15 ± 5% +4.8 17.99 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.00 +5.1 5.10 ± 3% perf-profile.calltrace.cycles-pp.__bitmap_weight.bitmap_bitremap.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages 0.00 +6.7 6.65 ± 4% perf-profile.calltrace.cycles-pp.__bitmap_weight.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64 0.00 +7.9 7.85 ± 4% perf-profile.calltrace.cycles-pp.bitmap_bitremap.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64 0.00 +8.3 8.31 ± 6% perf-profile.calltrace.cycles-pp.queue_pages_pte_range.__walk_page_range.walk_page_range.queue_pages_range.migrate_to_node 0.00 +9.4 9.39 ± 6% perf-profile.calltrace.cycles-pp.__walk_page_range.walk_page_range.queue_pages_range.migrate_to_node.do_migrate_pages 0.00 +10.3 10.28 ± 6% perf-profile.calltrace.cycles-pp.walk_page_range.queue_pages_range.migrate_to_node.do_migrate_pages.kernel_migrate_pages 0.00 +10.3 10.31 ± 6% perf-profile.calltrace.cycles-pp.queue_pages_range.migrate_to_node.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages 0.63 ± 6% +14.3 14.88 ± 4% perf-profile.calltrace.cycles-pp.migrate_to_node.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64 1.25 ± 3% +31.2 32.42 ± 4% perf-profile.calltrace.cycles-pp.do_migrate_pages.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.26 ± 3% +31.4 32.62 ± 4% perf-profile.calltrace.cycles-pp.__x64_sys_migrate_pages.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.26 ± 3% +31.4 32.62 ± 4% perf-profile.calltrace.cycles-pp.kernel_migrate_pages.__x64_sys_migrate_pages.do_syscall_64.entry_SYSCALL_64_after_hwframe 360536 ± 3% +1526.6% 5864634 ± 6% interrupts.CAL:Function_call_interrupts 4122 ± 31% +65.0% 6803 interrupts.CPU0.NMI:Non-maskable_interrupts 4122 ± 31% +65.0% 6803 interrupts.CPU0.PMI:Performance_monitoring_interrupts 6247 ± 9% +23.4% 7708 ± 17% interrupts.CPU0.RES:Rescheduling_interrupts 3735 ± 15% +2257.0% 88044 ± 33% interrupts.CPU1.CAL:Function_call_interrupts 6593 ± 16% +23.6% 8150 ± 11% interrupts.CPU1.RES:Rescheduling_interrupts 1605 ± 48% +5360.4% 87680 ± 33% interrupts.CPU1.TLB:TLB_shootdowns 3668 ± 27% +2395.9% 91551 ± 17% interrupts.CPU10.CAL:Function_call_interrupts 3705 ± 45% +87.8% 6957 ± 14% interrupts.CPU10.NMI:Non-maskable_interrupts 3705 ± 45% +87.8% 6957 ± 14% interrupts.CPU10.PMI:Performance_monitoring_interrupts 3692 ± 5% +50.9% 5571 ± 2% interrupts.CPU10.RES:Rescheduling_interrupts 1194 ± 74% +7530.5% 91146 ± 18% interrupts.CPU10.TLB:TLB_shootdowns 3833 ± 29% +2767.2% 109921 ± 10% interrupts.CPU11.CAL:Function_call_interrupts 3631 ± 2% +86.9% 6787 ± 8% interrupts.CPU11.RES:Rescheduling_interrupts 1569 ± 76% +6909.2% 109992 ± 11% interrupts.CPU11.TLB:TLB_shootdowns 3557 ± 35% +2148.0% 79971 ± 54% interrupts.CPU12.CAL:Function_call_interrupts 3760 ± 11% +58.0% 5942 ± 22% interrupts.CPU12.RES:Rescheduling_interrupts 1104 ±119% +7079.3% 79277 ± 56% interrupts.CPU12.TLB:TLB_shootdowns 3639 ± 32% +2257.8% 85817 ± 28% interrupts.CPU13.CAL:Function_call_interrupts 3497 ± 65% +76.1% 6159 ± 26% interrupts.CPU13.NMI:Non-maskable_interrupts 3497 ± 65% +76.1% 6159 ± 26% interrupts.CPU13.PMI:Performance_monitoring_interrupts 3799 ± 3% +46.3% 5558 ± 11% interrupts.CPU13.RES:Rescheduling_interrupts 1329 ± 89% +6373.1% 86075 ± 29% interrupts.CPU13.TLB:TLB_shootdowns 3942 ± 31% +2512.7% 103013 ± 12% interrupts.CPU14.CAL:Function_call_interrupts 3736 ± 4% +76.7% 6601 ± 14% interrupts.CPU14.RES:Rescheduling_interrupts 1491 ± 75% +6811.2% 103097 ± 12% interrupts.CPU14.TLB:TLB_shootdowns 3605 ± 33% +2212.0% 83354 ± 58% interrupts.CPU15.CAL:Function_call_interrupts 3624 ± 62% +85.4% 6720 ± 22% interrupts.CPU15.NMI:Non-maskable_interrupts 3624 ± 62% +85.4% 6720 ± 22% interrupts.CPU15.PMI:Performance_monitoring_interrupts 3751 ± 3% +41.9% 5325 ± 19% interrupts.CPU15.RES:Rescheduling_interrupts 1208 ± 98% +6788.3% 83227 ± 59% interrupts.CPU15.TLB:TLB_shootdowns 3738 ± 28% +2823.1% 109274 ± 29% interrupts.CPU16.CAL:Function_call_interrupts 3584 ± 3% +73.9% 6232 ± 17% interrupts.CPU16.RES:Rescheduling_interrupts 1749 ± 76% +6151.7% 109343 ± 29% interrupts.CPU16.TLB:TLB_shootdowns 3590 ± 36% +2164.9% 81327 ± 44% interrupts.CPU17.CAL:Function_call_interrupts 3515 ± 66% +67.5% 5887 ± 31% interrupts.CPU17.NMI:Non-maskable_interrupts 3515 ± 66% +67.5% 5887 ± 31% interrupts.CPU17.PMI:Performance_monitoring_interrupts 3602 ± 2% +59.5% 5746 ± 14% interrupts.CPU17.RES:Rescheduling_interrupts 1191 ±109% +6686.6% 80845 ± 46% interrupts.CPU17.TLB:TLB_shootdowns 4548 ± 23% +2020.8% 96469 ± 48% interrupts.CPU18.CAL:Function_call_interrupts 3822 ± 8% +52.3% 5822 ± 21% interrupts.CPU18.RES:Rescheduling_interrupts 2195 ± 50% +4291.9% 96414 ± 48% interrupts.CPU18.TLB:TLB_shootdowns 3596 ± 15% +1942.4% 73450 ± 82% interrupts.CPU2.CAL:Function_call_interrupts 5367 ± 6% +33.2% 7150 ± 6% interrupts.CPU2.NMI:Non-maskable_interrupts 5367 ± 6% +33.2% 7150 ± 6% interrupts.CPU2.PMI:Performance_monitoring_interrupts 3608 ± 33% +1494.0% 57512 ± 48% interrupts.CPU20.CAL:Function_call_interrupts 3723 ± 3% +36.6% 5085 ± 11% interrupts.CPU20.RES:Rescheduling_interrupts 1159 ± 97% +4754.8% 56304 ± 50% interrupts.CPU20.TLB:TLB_shootdowns 3686 ± 30% +2621.0% 100315 ± 33% interrupts.CPU21.CAL:Function_call_interrupts 3574 ± 67% +77.6% 6349 ± 25% interrupts.CPU21.NMI:Non-maskable_interrupts 3574 ± 67% +77.6% 6349 ± 25% interrupts.CPU21.PMI:Performance_monitoring_interrupts 3835 ± 13% +65.2% 6337 ± 15% interrupts.CPU21.RES:Rescheduling_interrupts 1162 ± 98% +8509.6% 100065 ± 34% interrupts.CPU21.TLB:TLB_shootdowns 3687 ± 27% +2336.0% 89813 ± 50% interrupts.CPU22.CAL:Function_call_interrupts 3810 +83.5% 6993 ± 27% interrupts.CPU22.RES:Rescheduling_interrupts 1484 ± 66% +5953.7% 89867 ± 51% interrupts.CPU22.TLB:TLB_shootdowns 3974 ± 17% +2750.3% 113270 ± 15% interrupts.CPU23.CAL:Function_call_interrupts 3926 ± 11% +80.7% 7096 ± 9% interrupts.CPU23.RES:Rescheduling_interrupts 1464 ± 53% +7676.5% 113848 ± 15% interrupts.CPU23.TLB:TLB_shootdowns 4345 ± 14% +2287.9% 103770 ± 13% interrupts.CPU24.CAL:Function_call_interrupts 4072 ± 9% +84.4% 7507 ± 13% interrupts.CPU24.RES:Rescheduling_interrupts 1858 ± 38% +5495.7% 103967 ± 14% interrupts.CPU24.TLB:TLB_shootdowns 3807 ± 26% +2463.3% 97585 ± 30% interrupts.CPU25.CAL:Function_call_interrupts 3852 ± 4% +78.3% 6870 ± 14% interrupts.CPU25.RES:Rescheduling_interrupts 1368 ± 72% +7011.0% 97296 ± 31% interrupts.CPU25.TLB:TLB_shootdowns 3574 ± 21% +1729.3% 65390 ± 48% interrupts.CPU26.CAL:Function_call_interrupts 6245 ± 7% -30.6% 4334 ± 26% interrupts.CPU26.NMI:Non-maskable_interrupts 6245 ± 7% -30.6% 4334 ± 26% interrupts.CPU26.PMI:Performance_monitoring_interrupts 4092 ± 18% +37.4% 5623 ± 14% interrupts.CPU26.RES:Rescheduling_interrupts 1168 ± 64% +5429.2% 64594 ± 50% interrupts.CPU26.TLB:TLB_shootdowns 6377 ± 10% -34.5% 4179 ± 23% interrupts.CPU27.NMI:Non-maskable_interrupts 6377 ± 10% -34.5% 4179 ± 23% interrupts.CPU27.PMI:Performance_monitoring_interrupts 6425 ± 14% -38.7% 3937 ± 23% interrupts.CPU28.NMI:Non-maskable_interrupts 6425 ± 14% -38.7% 3937 ± 23% interrupts.CPU28.PMI:Performance_monitoring_interrupts 4880 ± 4% +894.4% 48528 ± 19% interrupts.CPU29.CAL:Function_call_interrupts 3951 ± 6% +25.6% 4962 ± 6% interrupts.CPU29.RES:Rescheduling_interrupts 2365 ± 13% +1898.5% 47274 ± 20% interrupts.CPU29.TLB:TLB_shootdowns 3890 ± 13% +3046.4% 122404 ± 11% interrupts.CPU3.CAL:Function_call_interrupts 4834 ± 22% +53.2% 7407 ± 4% interrupts.CPU3.NMI:Non-maskable_interrupts 4834 ± 22% +53.2% 7407 ± 4% interrupts.CPU3.PMI:Performance_monitoring_interrupts 4600 ± 3% +89.1% 8699 ± 25% interrupts.CPU3.RES:Rescheduling_interrupts 1585 ± 50% +7626.2% 122460 ± 12% interrupts.CPU3.TLB:TLB_shootdowns 4598 ± 15% +1179.7% 58846 ± 45% interrupts.CPU31.CAL:Function_call_interrupts 3672 ± 3% +61.8% 5941 ± 14% interrupts.CPU31.RES:Rescheduling_interrupts 2274 ± 27% +2438.3% 57732 ± 47% interrupts.CPU31.TLB:TLB_shootdowns 4554 ± 21% +1114.9% 55329 ± 56% interrupts.CPU32.CAL:Function_call_interrupts 3741 +44.6% 5408 ± 21% interrupts.CPU32.RES:Rescheduling_interrupts 2105 ± 41% +2460.0% 53894 ± 59% interrupts.CPU32.TLB:TLB_shootdowns 3907 ± 30% +860.8% 37539 ± 65% interrupts.CPU35.CAL:Function_call_interrupts 3871 ± 9% +22.8% 4752 ± 18% interrupts.CPU35.RES:Rescheduling_interrupts 1638 ± 73% +2093.4% 35938 ± 68% interrupts.CPU35.TLB:TLB_shootdowns 4647 ± 23% +1627.4% 80281 ± 12% interrupts.CPU36.CAL:Function_call_interrupts 3821 ± 5% +57.4% 6015 ± 9% interrupts.CPU36.RES:Rescheduling_interrupts 2181 ± 45% +3557.4% 79776 ± 12% interrupts.CPU36.TLB:TLB_shootdowns 7070 ± 14% -30.3% 4924 ± 24% interrupts.CPU38.NMI:Non-maskable_interrupts 7070 ± 14% -30.3% 4924 ± 24% interrupts.CPU38.PMI:Performance_monitoring_interrupts 3711 ± 12% +1962.2% 76543 ± 53% interrupts.CPU4.CAL:Function_call_interrupts 4987 ± 21% +43.4% 7152 ± 3% interrupts.CPU4.NMI:Non-maskable_interrupts 4987 ± 21% +43.4% 7152 ± 3% interrupts.CPU4.PMI:Performance_monitoring_interrupts 3815 ± 2% +60.4% 6121 ± 20% interrupts.CPU4.RES:Rescheduling_interrupts 1324 ± 33% +5629.5% 75872 ± 55% interrupts.CPU4.TLB:TLB_shootdowns 4648 ± 22% +814.8% 42517 ± 39% interrupts.CPU43.CAL:Function_call_interrupts 2282 ± 35% +1704.8% 41186 ± 41% interrupts.CPU43.TLB:TLB_shootdowns 3602 ± 29% +2533.9% 94873 ± 42% interrupts.CPU45.CAL:Function_call_interrupts 4435 ± 44% +62.7% 7217 ± 10% interrupts.CPU45.NMI:Non-maskable_interrupts 4435 ± 44% +62.7% 7217 ± 10% interrupts.CPU45.PMI:Performance_monitoring_interrupts 3712 ± 4% +57.8% 5856 ± 21% interrupts.CPU45.RES:Rescheduling_interrupts 1180 ± 98% +7913.7% 94601 ± 44% interrupts.CPU45.TLB:TLB_shootdowns 4762 ± 26% +49.6% 7124 ± 12% interrupts.CPU46.NMI:Non-maskable_interrupts 4762 ± 26% +49.6% 7124 ± 12% interrupts.CPU46.PMI:Performance_monitoring_interrupts 1165 ±115% +5808.7% 68836 ± 84% interrupts.CPU47.TLB:TLB_shootdowns 3486 ± 30% +2453.1% 89006 ± 60% interrupts.CPU48.CAL:Function_call_interrupts 4565 ± 46% +55.6% 7104 ± 14% interrupts.CPU48.NMI:Non-maskable_interrupts 4565 ± 46% +55.6% 7104 ± 14% interrupts.CPU48.PMI:Performance_monitoring_interrupts 3695 ± 2% +70.3% 6295 ± 20% interrupts.CPU48.RES:Rescheduling_interrupts 1073 ±111% +8148.9% 88511 ± 62% interrupts.CPU48.TLB:TLB_shootdowns 4106 ± 47% +69.1% 6944 ± 14% interrupts.CPU49.NMI:Non-maskable_interrupts 4106 ± 47% +69.1% 6944 ± 14% interrupts.CPU49.PMI:Performance_monitoring_interrupts 1185 ± 95% +5391.3% 65112 ± 98% interrupts.CPU49.TLB:TLB_shootdowns 3601 ± 8% +2568.7% 96098 ± 29% interrupts.CPU5.CAL:Function_call_interrupts 3785 ± 26% +91.1% 7233 ± 3% interrupts.CPU5.NMI:Non-maskable_interrupts 3785 ± 26% +91.1% 7233 ± 3% interrupts.CPU5.PMI:Performance_monitoring_interrupts 3668 ± 2% +74.1% 6386 ± 16% interrupts.CPU5.RES:Rescheduling_interrupts 1137 ± 29% +8322.2% 95802 ± 30% interrupts.CPU5.TLB:TLB_shootdowns 3745 ± 41% +84.0% 6890 ± 18% interrupts.CPU50.NMI:Non-maskable_interrupts 3745 ± 41% +84.0% 6890 ± 18% interrupts.CPU50.PMI:Performance_monitoring_interrupts 654.50 ± 68% +12823.4% 84583 ± 66% interrupts.CPU50.TLB:TLB_shootdowns 3754 ± 59% +84.3% 6921 ± 18% interrupts.CPU51.NMI:Non-maskable_interrupts 3754 ± 59% +84.3% 6921 ± 18% interrupts.CPU51.PMI:Performance_monitoring_interrupts 3616 ± 4% +63.4% 5907 ± 23% interrupts.CPU51.RES:Rescheduling_interrupts 1051 ±115% +7965.0% 84783 ± 67% interrupts.CPU51.TLB:TLB_shootdowns 3537 ± 35% +1280.8% 48845 ± 62% interrupts.CPU52.CAL:Function_call_interrupts 1067 ±112% +4353.9% 47545 ± 65% interrupts.CPU52.TLB:TLB_shootdowns 1112 ±109% +4960.5% 56285 ± 97% interrupts.CPU53.TLB:TLB_shootdowns 3841 ± 56% +80.5% 6932 ± 20% interrupts.CPU54.NMI:Non-maskable_interrupts 3841 ± 56% +80.5% 6932 ± 20% interrupts.CPU54.PMI:Performance_monitoring_interrupts 3561 +69.5% 6036 ± 23% interrupts.CPU54.RES:Rescheduling_interrupts 3543 ± 32% +1357.6% 51648 ± 60% interrupts.CPU55.CAL:Function_call_interrupts 3537 ± 2% +39.5% 4933 ± 13% interrupts.CPU55.RES:Rescheduling_interrupts 1144 ±103% +4304.4% 50397 ± 63% interrupts.CPU55.TLB:TLB_shootdowns 3550 ± 35% +2444.4% 90343 ± 56% interrupts.CPU56.CAL:Function_call_interrupts 3650 ± 65% +94.0% 7082 ± 11% interrupts.CPU56.NMI:Non-maskable_interrupts 3650 ± 65% +94.0% 7082 ± 11% interrupts.CPU56.PMI:Performance_monitoring_interrupts 3651 +65.0% 6023 ± 23% interrupts.CPU56.RES:Rescheduling_interrupts 1136 ±112% +7824.5% 90062 ± 58% interrupts.CPU56.TLB:TLB_shootdowns 2575 ± 33% +170.5% 6966 ± 17% interrupts.CPU57.NMI:Non-maskable_interrupts 2575 ± 33% +170.5% 6966 ± 17% interrupts.CPU57.PMI:Performance_monitoring_interrupts 3533 ± 3% +63.2% 5768 ± 24% interrupts.CPU57.RES:Rescheduling_interrupts 1119 ±104% +7722.0% 87547 ± 60% interrupts.CPU57.TLB:TLB_shootdowns 3626 +52.3% 5525 ± 18% interrupts.CPU58.RES:Rescheduling_interrupts 3579 ± 35% +2484.6% 92516 ± 52% interrupts.CPU59.CAL:Function_call_interrupts 2676 ± 32% +152.7% 6762 ± 23% interrupts.CPU59.NMI:Non-maskable_interrupts 2676 ± 32% +152.7% 6762 ± 23% interrupts.CPU59.PMI:Performance_monitoring_interrupts 3615 +63.4% 5907 ± 22% interrupts.CPU59.RES:Rescheduling_interrupts 1144 ±104% +7945.0% 92075 ± 53% interrupts.CPU59.TLB:TLB_shootdowns 3821 ± 15% +1528.2% 62212 ± 38% interrupts.CPU6.CAL:Function_call_interrupts 3942 ± 39% +71.2% 6748 ± 9% interrupts.CPU6.NMI:Non-maskable_interrupts 3942 ± 39% +71.2% 6748 ± 9% interrupts.CPU6.PMI:Performance_monitoring_interrupts 3903 ± 4% +47.5% 5759 ± 22% interrupts.CPU6.RES:Rescheduling_interrupts 1496 ± 36% +3987.9% 61185 ± 39% interrupts.CPU6.TLB:TLB_shootdowns 2798 ± 22% +139.4% 6699 ± 21% interrupts.CPU60.NMI:Non-maskable_interrupts 2798 ± 22% +139.4% 6699 ± 21% interrupts.CPU60.PMI:Performance_monitoring_interrupts 2957 ± 17% +123.3% 6604 ± 22% interrupts.CPU61.NMI:Non-maskable_interrupts 2957 ± 17% +123.3% 6604 ± 22% interrupts.CPU61.PMI:Performance_monitoring_interrupts 3680 ± 37% +2380.7% 91308 ± 40% interrupts.CPU62.CAL:Function_call_interrupts 4022 ± 14% +37.9% 5547 ± 11% interrupts.CPU62.RES:Rescheduling_interrupts 1136 ±124% +7892.6% 90855 ± 42% interrupts.CPU62.TLB:TLB_shootdowns 3634 ± 36% +2919.3% 109735 ± 24% interrupts.CPU63.CAL:Function_call_interrupts 3521 ± 26% +114.0% 7538 ± 2% interrupts.CPU63.NMI:Non-maskable_interrupts 3521 ± 26% +114.0% 7538 ± 2% interrupts.CPU63.PMI:Performance_monitoring_interrupts 3948 ± 10% +65.9% 6551 ± 18% interrupts.CPU63.RES:Rescheduling_interrupts 1152 ±110% +9401.5% 109528 ± 24% interrupts.CPU63.TLB:TLB_shootdowns 3592 ± 34% +3034.8% 112602 ± 26% interrupts.CPU64.CAL:Function_call_interrupts 2961 ± 17% +148.8% 7369 ± 9% interrupts.CPU64.NMI:Non-maskable_interrupts 2961 ± 17% +148.8% 7369 ± 9% interrupts.CPU64.PMI:Performance_monitoring_interrupts 4193 ± 19% +66.8% 6996 ± 10% interrupts.CPU64.RES:Rescheduling_interrupts 1104 ±106% +10101.5% 112650 ± 27% interrupts.CPU64.TLB:TLB_shootdowns 3509 ± 33% +2502.5% 91320 ± 30% interrupts.CPU65.CAL:Function_call_interrupts 2965 ± 17% +145.6% 7283 ± 6% interrupts.CPU65.NMI:Non-maskable_interrupts 2965 ± 17% +145.6% 7283 ± 6% interrupts.CPU65.PMI:Performance_monitoring_interrupts 3859 ± 8% +52.9% 5899 ± 13% interrupts.CPU65.RES:Rescheduling_interrupts 1037 ±112% +8669.9% 91009 ± 31% interrupts.CPU65.TLB:TLB_shootdowns 3370 ± 10% +2036.3% 71994 ± 5% interrupts.CPU7.CAL:Function_call_interrupts 3469 ± 44% +72.9% 5999 ± 24% interrupts.CPU7.NMI:Non-maskable_interrupts 3469 ± 44% +72.9% 5999 ± 24% interrupts.CPU7.PMI:Performance_monitoring_interrupts 3809 ± 4% +47.7% 5625 ± 11% interrupts.CPU7.RES:Rescheduling_interrupts 988.50 ± 61% +7115.2% 71322 ± 6% interrupts.CPU7.TLB:TLB_shootdowns 3654 ± 4% +80.9% 6612 ± 24% interrupts.CPU77.RES:Rescheduling_interrupts 3569 ± 25% +2737.1% 101271 ± 22% interrupts.CPU8.CAL:Function_call_interrupts 3412 ± 47% +81.7% 6202 ± 26% interrupts.CPU8.NMI:Non-maskable_interrupts 3412 ± 47% +81.7% 6202 ± 26% interrupts.CPU8.PMI:Performance_monitoring_interrupts 3822 ± 6% +72.8% 6605 ± 13% interrupts.CPU8.RES:Rescheduling_interrupts 2371 ± 57% +4169.2% 101244 ± 23% interrupts.CPU8.TLB:TLB_shootdowns 3280 ± 25% +2403.5% 82113 ± 56% interrupts.CPU9.CAL:Function_call_interrupts 3406 ± 50% +101.8% 6873 ± 11% interrupts.CPU9.NMI:Non-maskable_interrupts 3406 ± 50% +101.8% 6873 ± 11% interrupts.CPU9.PMI:Performance_monitoring_interrupts 3636 ± 6% +66.1% 6041 ± 24% interrupts.CPU9.RES:Rescheduling_interrupts 898.75 ±100% +8966.6% 81486 ± 58% interrupts.CPU9.TLB:TLB_shootdowns 441700 ± 6% +18.5% 523360 ± 5% interrupts.NMI:Non-maskable_interrupts 441700 ± 6% +18.5% 523360 ± 5% interrupts.PMI:Performance_monitoring_interrupts 344450 ± 2% +42.8% 491951 ± 2% interrupts.RES:Rescheduling_interrupts 153414 ± 10% +3677.4% 5795118 ± 6% interrupts.TLB:TLB_shootdowns stress-ng.numa.ops 3500 +-+------------------------------------------------------------------+ | O O | 3000 +-+ O O O OO O O | O O O O O O O O OO | 2500 +-+ O O | | O O O O OO O OO | 2000 +-+ O O | | | 1500 +-+ | | | 1000 +-+ | | | 500 +-+ | | | 0 +-+------------------------------------------------------------------+ stress-ng.numa.ops_per_sec 700 +-+-------------------------------------------------------------------+ | O | 600 +-+ O O O OO O O O | O O O O O O O O O O | 500 +-+ O | | O OO OO O O OO O | 400 +-+ O O | | | 300 +-+ | | | 200 +-+ | | | 100 +-+ | | | 0 +-+-------------------------------------------------------------------+ [*] bisect-good sample [O] bisect-bad sample *************************************************************************************************** vm-snb-8G: qemu-system-x86_64 -enable-kvm -cpu SandyBridge -smp 2 -m 8G ========================================================================================= compiler/kconfig/rootfs/tbox_group/test/testcase: gcc-7/x86_64-rhel-7.2/debian-x86_64-2018-04-03.cgz/vm-snb-8G/cve/ltp commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | :12 8% 1:12 dmesg.BUG:soft_lockup-CPU##stuck_for#s :12 83% 10:12 dmesg.Kernel_panic-not_syncing:System_is_deadlocked_on_memory :12 8% 1:12 dmesg.Kernel_panic-not_syncing:softlockup:hung_tasks :12 83% 10:12 dmesg.Out_of_memory:Kill_process :12 83% 10:12 dmesg.Out_of_memory_and_no_killable_processes :12 8% 1:12 dmesg.RIP:free_reserved_area %stddev %change %stddev \ | \ 1.00 -100.0% 0.00 ltp.cve-2017-18075.pass 783.53 ± 40% +165.9% 2083 ltp.time.elapsed_time 783.53 ± 40% +165.9% 2083 ltp.time.elapsed_time.max 1171525 ± 21% -69.1% 362268 ltp.time.involuntary_context_switches 11374889 ± 27% -89.7% 1171686 ltp.time.minor_page_faults 103.08 ± 9% -46.6% 55.00 ltp.time.percent_of_cpu_this_job_got 199802 ± 12% -36.0% 127912 ltp.time.voluntary_context_switches 557835 ± 3% +469.2% 3175044 meminfo.Active 557602 ± 3% +469.4% 3174826 meminfo.Active(anon) 204797 -51.1% 100217 meminfo.CmaFree 709192 ± 3% -11.8% 625168 meminfo.Inactive 22617 ± 16% -50.8% 11135 meminfo.Inactive(anon) 4772734 -52.9% 2247863 meminfo.MemAvailable 4273603 -57.4% 1820264 meminfo.MemFree 3888307 +63.1% 6341643 meminfo.Memused *************************************************************************************************** lkp-skl-4sp1: 192 threads Skylake-4S with 704G memory ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: gcc-7/performance/x86_64-rhel-7.2/process/100%/debian-x86_64-2018-04-03.cgz/lkp-skl-4sp1/futex1/will-it-scale commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- %stddev %change %stddev \ | \ 0.42 ± 41% -37.6% 0.27 ± 38% turbostat.CPU%c1 436529 +18.1% 515400 ± 2% meminfo.Active 432417 +18.2% 511272 ± 2% meminfo.Active(anon) 1.01 ±100% +1140.2% 12.50 ±141% irq_exception_noise.__do_page_fault.50th 1.10 ±100% +1059.7% 12.70 ±137% irq_exception_noise.__do_page_fault.60th 1.18 ±100% +1002.0% 13.00 ±133% irq_exception_noise.__do_page_fault.70th 1.32 ±100% +934.8% 13.71 ±123% irq_exception_noise.__do_page_fault.80th 545804 ± 35% -56.4% 238144 ± 29% numa-numastat.node0.local_node 568271 ± 32% -54.7% 257664 ± 24% numa-numastat.node0.numa_hit 232178 ± 44% +135.2% 546083 ± 30% numa-numastat.node2.local_node 268049 ± 38% +114.9% 576076 ± 27% numa-numastat.node2.numa_hit 330370 ± 19% -39.1% 201131 ± 50% numa-numastat.node3.local_node 342159 ± 15% -31.7% 233865 ± 42% numa-numastat.node3.numa_hit 108167 +18.2% 127807 ± 2% proc-vmstat.nr_active_anon 6541 -1.8% 6425 proc-vmstat.nr_inactive_anon 8474 -2.4% 8271 ± 2% proc-vmstat.nr_mapped 108167 +18.2% 127807 ± 2% proc-vmstat.nr_zone_active_anon 6541 -1.8% 6425 proc-vmstat.nr_zone_inactive_anon 29259 +3.5% 30283 proc-vmstat.pgactivate 1477807 -2.1% 1446297 proc-vmstat.pgfree 13.29 ± 5% -2.6 10.72 ± 4% perf-stat.i.cache-miss-rate% 1107530 ± 4% -16.6% 923800 ± 3% perf-stat.i.cache-misses 57.31 -2.0% 56.14 perf-stat.i.cpu-migrations 429491 ± 3% +14.9% 493355 ± 4% perf-stat.i.cycles-between-cache-misses 51.31 +10.9 62.22 ± 6% perf-stat.i.iTLB-load-miss-rate% 3.369e+08 +2.2% 3.443e+08 perf-stat.i.iTLB-load-misses 3.2e+08 -32.8% 2.149e+08 ± 16% perf-stat.i.iTLB-loads 863.04 -2.4% 842.12 perf-stat.i.instructions-per-iTLB-miss 355.57 ± 6% -13.9% 306.21 ± 5% sched_debug.cfs_rq:/.exec_clock.stddev 0.46 +163.6% 1.21 ± 10% sched_debug.cfs_rq:/.nr_spread_over.avg 619.50 ± 6% -13.9% 533.25 ± 7% sched_debug.cfs_rq:/.util_est_enqueued.max 128891 +42.6% 183838 ± 30% sched_debug.cpu.avg_idle.stddev 11367 ± 2% +14.6% 13028 sched_debug.cpu.curr->pid.max 2547 ± 15% +30.3% 3321 sched_debug.cpu.curr->pid.min -11.50 -23.9% -8.75 sched_debug.cpu.nr_uninterruptible.min 10647 ± 20% -40.5% 6331 ± 17% sched_debug.cpu.sched_goidle.max 800.32 ± 19% -31.8% 545.75 ± 5% sched_debug.cpu.sched_goidle.stddev 20627 +16.3% 23980 ± 6% softirqs.CPU0.RCU 112466 ± 4% -8.4% 102965 ± 4% softirqs.CPU0.TIMER 20826 +10.9% 23088 ± 6% softirqs.CPU110.RCU 20775 ± 8% +22.1% 25361 ± 5% softirqs.CPU112.RCU 20835 ± 10% +22.8% 25592 ± 5% softirqs.CPU113.RCU 21208 ± 6% +21.7% 25811 ± 4% softirqs.CPU114.RCU 20691 ± 7% +24.5% 25764 ± 6% softirqs.CPU115.RCU 20809 ± 8% +22.8% 25550 ± 5% softirqs.CPU116.RCU 20870 ± 9% +23.4% 25754 ± 4% softirqs.CPU117.RCU 20714 ± 8% +23.4% 25568 ± 5% softirqs.CPU118.RCU 20652 ± 8% +22.6% 25310 ± 5% softirqs.CPU119.RCU 20909 +10.8% 23171 ± 7% softirqs.CPU13.RCU 20488 +10.1% 22548 ± 5% softirqs.CPU14.RCU 20955 ± 6% +21.3% 25421 ± 6% softirqs.CPU16.RCU 19946 +15.2% 22982 ± 11% softirqs.CPU173.RCU 133691 ± 19% -22.1% 104181 ± 4% softirqs.CPU176.TIMER 21426 ± 3% +17.6% 25194 ± 4% softirqs.CPU18.RCU 20370 ± 6% +24.1% 25275 ± 3% softirqs.CPU19.RCU 21029 ± 4% +19.6% 25146 ± 5% softirqs.CPU20.RCU 20971 ± 8% +19.7% 25095 ± 4% softirqs.CPU21.RCU 20759 ± 6% +20.2% 24958 ± 5% softirqs.CPU22.RCU 20627 ± 6% +21.6% 25077 ± 5% softirqs.CPU23.RCU 20693 +10.5% 22867 ± 6% softirqs.CPU5.RCU 18951 ± 4% +21.1% 22953 ± 12% softirqs.CPU75.RCU 110575 ± 4% -7.5% 102329 ± 2% softirqs.CPU98.TIMER 57149 +37.8% 78776 ± 10% numa-vmstat.node0.nr_file_pages 1470 +84.0% 2706 ± 37% numa-vmstat.node0.nr_mapped 1020 ± 74% +1687.7% 18243 ± 45% numa-vmstat.node0.nr_shmem 16.50 ± 33% +8503.0% 1419 ±122% numa-vmstat.node1.nr_inactive_anon 6205 +19.9% 7437 ± 10% numa-vmstat.node1.nr_kernel_stack 594.50 ± 3% +141.9% 1438 ± 52% numa-vmstat.node1.nr_page_table_pages 56.00 ± 33% +2705.4% 1571 ±112% numa-vmstat.node1.nr_shmem 3072 ± 5% +119.8% 6751 ± 17% numa-vmstat.node1.nr_slab_reclaimable 11562 ± 2% +31.6% 15218 ± 9% numa-vmstat.node1.nr_slab_unreclaimable 16.50 ± 33% +8503.0% 1419 ±122% numa-vmstat.node1.nr_zone_inactive_anon 266341 ± 12% +131.7% 617001 ± 9% numa-vmstat.node1.numa_hit 140430 ± 23% +258.4% 503291 ± 12% numa-vmstat.node1.numa_local 1028 -100.0% 0.00 numa-vmstat.node2.nr_active_file 71145 -19.0% 57616 ± 2% numa-vmstat.node2.nr_file_pages 4668 ± 12% -67.9% 1500 ±104% numa-vmstat.node2.nr_inactive_anon 544.50 -100.0% 0.00 numa-vmstat.node2.nr_inactive_file 9199 ± 4% -28.3% 6593 ± 5% numa-vmstat.node2.nr_kernel_stack 3552 -50.5% 1758 ± 33% numa-vmstat.node2.nr_mapped 2955 ± 6% -65.5% 1018 ± 38% numa-vmstat.node2.nr_page_table_pages 4996 ± 11% -65.4% 1727 ± 95% numa-vmstat.node2.nr_shmem 64607 -13.5% 55888 numa-vmstat.node2.nr_unevictable 1028 -100.0% 0.00 numa-vmstat.node2.nr_zone_active_file 4668 ± 12% -67.9% 1500 ±104% numa-vmstat.node2.nr_zone_inactive_anon 544.50 -100.0% 0.00 numa-vmstat.node2.nr_zone_inactive_file 64607 -13.5% 55888 numa-vmstat.node2.nr_zone_unevictable 31367 ± 22% -70.8% 9145 ±104% numa-vmstat.node3.nr_active_anon 11913 ± 59% -87.6% 1476 ± 52% numa-vmstat.node3.nr_anon_pages 1000 ± 17% -72.0% 279.75 ±150% numa-vmstat.node3.nr_inactive_anon 698.00 ± 11% -15.7% 588.25 numa-vmstat.node3.nr_page_table_pages 6430 ± 14% -39.1% 3918 ± 20% numa-vmstat.node3.nr_slab_reclaimable 31366 ± 22% -70.8% 9145 ±104% numa-vmstat.node3.nr_zone_active_anon 1000 ± 17% -72.0% 279.75 ±150% numa-vmstat.node3.nr_zone_inactive_anon 345814 ± 42% -45.4% 188830 ± 51% numa-vmstat.node3.numa_local 228598 +37.8% 315090 ± 10% numa-meminfo.node0.FilePages 3152 ± 89% +354.3% 14320 ± 50% numa-meminfo.node0.Inactive 5883 +80.6% 10625 ± 34% numa-meminfo.node0.Mapped 4082 ± 74% +1687.1% 72956 ± 45% numa-meminfo.node0.Shmem 68.50 ± 31% +8993.8% 6229 ±107% numa-meminfo.node1.Inactive 68.50 ± 31% +8191.2% 5679 ±122% numa-meminfo.node1.Inactive(anon) 12286 ± 5% +119.8% 27005 ± 17% numa-meminfo.node1.KReclaimable 6207 +19.9% 7441 ± 10% numa-meminfo.node1.KernelStack 2372 ± 3% +142.7% 5757 ± 52% numa-meminfo.node1.PageTables 12286 ± 5% +119.8% 27005 ± 17% numa-meminfo.node1.SReclaimable 46242 ± 2% +31.6% 60873 ± 9% numa-meminfo.node1.SUnreclaim 223.50 ± 33% +2711.9% 6284 ±112% numa-meminfo.node1.Shmem 58528 +50.1% 87879 ± 11% numa-meminfo.node1.Slab 4112 -100.0% 0.00 numa-meminfo.node2.Active(file) 284580 -19.0% 230466 ± 2% numa-meminfo.node2.FilePages 20854 ± 11% -71.2% 6004 ±104% numa-meminfo.node2.Inactive 18673 ± 12% -67.8% 6004 ±104% numa-meminfo.node2.Inactive(anon) 2180 -100.0% 0.00 numa-meminfo.node2.Inactive(file) 9203 ± 4% -28.3% 6594 ± 5% numa-meminfo.node2.KernelStack 13712 -48.6% 7054 ± 33% numa-meminfo.node2.Mapped 883222 ± 6% -16.2% 740514 ± 13% numa-meminfo.node2.MemUsed 11829 ± 6% -65.8% 4040 ± 38% numa-meminfo.node2.PageTables 19985 ± 11% -65.4% 6911 ± 95% numa-meminfo.node2.Shmem 258431 -13.5% 223554 numa-meminfo.node2.Unevictable 125256 ± 22% -70.8% 36547 ±104% numa-meminfo.node3.Active 125256 ± 22% -70.8% 36547 ±104% numa-meminfo.node3.Active(anon) 24571 ± 83% -96.0% 974.75 ±173% numa-meminfo.node3.AnonHugePages 47677 ± 59% -87.7% 5876 ± 52% numa-meminfo.node3.AnonPages 3993 ± 18% -72.1% 1115 ±149% numa-meminfo.node3.Inactive 3991 ± 18% -72.1% 1115 ±149% numa-meminfo.node3.Inactive(anon) 25714 ± 14% -39.1% 15671 ± 20% numa-meminfo.node3.KReclaimable 696078 ± 2% -14.9% 592511 ± 8% numa-meminfo.node3.MemUsed 2772 ± 11% -15.9% 2332 numa-meminfo.node3.PageTables 25714 ± 14% -39.1% 15671 ± 20% numa-meminfo.node3.SReclaimable 78860 ± 8% -16.7% 65670 ± 6% numa-meminfo.node3.Slab 1702 ± 25% -44.9% 937.50 ± 17% interrupts.CPU0.RES:Rescheduling_interrupts 6102 ± 5% -3.6% 5884 ± 5% interrupts.CPU100.CAL:Function_call_interrupts 6091 ± 5% -4.1% 5842 ± 5% interrupts.CPU101.CAL:Function_call_interrupts 6100 ± 5% -4.3% 5841 ± 6% interrupts.CPU102.CAL:Function_call_interrupts 6099 ± 5% -3.6% 5882 ± 5% interrupts.CPU103.CAL:Function_call_interrupts 6099 ± 5% -3.5% 5883 ± 5% interrupts.CPU104.CAL:Function_call_interrupts 6099 ± 5% -3.6% 5882 ± 5% interrupts.CPU106.CAL:Function_call_interrupts 6119 ± 5% -3.8% 5886 ± 5% interrupts.CPU107.CAL:Function_call_interrupts 6115 ± 5% -8.8% 5575 ± 4% interrupts.CPU108.CAL:Function_call_interrupts 6115 ± 5% -3.6% 5894 ± 4% interrupts.CPU109.CAL:Function_call_interrupts 6118 ± 5% -3.9% 5879 ± 5% interrupts.CPU110.CAL:Function_call_interrupts 6117 ± 5% -3.8% 5884 ± 4% interrupts.CPU111.CAL:Function_call_interrupts 291.50 ± 86% -90.1% 29.00 ± 33% interrupts.CPU111.RES:Rescheduling_interrupts 6117 ± 5% -3.6% 5897 ± 5% interrupts.CPU112.CAL:Function_call_interrupts 6117 ± 5% -3.5% 5902 ± 5% interrupts.CPU113.CAL:Function_call_interrupts 6116 ± 5% -3.5% 5904 ± 5% interrupts.CPU114.CAL:Function_call_interrupts 6116 ± 5% -3.6% 5897 ± 5% interrupts.CPU115.CAL:Function_call_interrupts 6117 ± 5% -3.6% 5897 ± 5% interrupts.CPU116.CAL:Function_call_interrupts 6118 ± 5% -3.6% 5896 ± 5% interrupts.CPU117.CAL:Function_call_interrupts 6088 ± 5% -5.7% 5744 interrupts.CPU12.CAL:Function_call_interrupts 1079 ± 78% -77.3% 245.50 ±134% interrupts.CPU120.RES:Rescheduling_interrupts 949.50 ± 91% -93.4% 63.00 ± 35% interrupts.CPU122.RES:Rescheduling_interrupts 6099 ± 5% -3.6% 5879 ± 4% interrupts.CPU13.CAL:Function_call_interrupts 2647 ± 18% -82.0% 477.00 ± 33% interrupts.CPU13.RES:Rescheduling_interrupts 208.50 ± 51% -51.3% 101.50 ± 36% interrupts.CPU14.RES:Rescheduling_interrupts 53.00 ± 50% +502.4% 319.25 ± 68% interrupts.CPU140.RES:Rescheduling_interrupts 20.00 ± 55% +392.5% 98.50 ± 35% interrupts.CPU141.RES:Rescheduling_interrupts 6093 ± 5% -4.2% 5838 ± 5% interrupts.CPU144.CAL:Function_call_interrupts 206.00 ± 59% -63.7% 74.75 ± 75% interrupts.CPU145.RES:Rescheduling_interrupts 6100 ± 5% -3.9% 5865 ± 5% interrupts.CPU15.CAL:Function_call_interrupts 8.00 +575.0% 54.00 ± 60% interrupts.CPU152.RES:Rescheduling_interrupts 969.00 ± 67% -91.8% 79.25 ± 56% interrupts.CPU153.RES:Rescheduling_interrupts 922.00 ± 91% -74.7% 233.50 ±134% interrupts.CPU159.RES:Rescheduling_interrupts 6098 ± 5% -3.9% 5863 ± 5% interrupts.CPU16.CAL:Function_call_interrupts 6097 ± 5% -4.2% 5841 ± 4% interrupts.CPU17.CAL:Function_call_interrupts 1077 ± 89% -94.5% 58.75 ± 78% interrupts.CPU176.RES:Rescheduling_interrupts 710144 ± 12% -13.5% 614236 interrupts.CPU182.LOC:Local_timer_interrupts 303.50 ± 78% -92.6% 22.50 ± 50% interrupts.CPU186.RES:Rescheduling_interrupts 6096 ± 5% -6.4% 5704 interrupts.CPU19.CAL:Function_call_interrupts 271.00 ± 71% -86.6% 36.25 ± 81% interrupts.CPU190.RES:Rescheduling_interrupts 6109 ± 5% -3.8% 5875 ± 5% interrupts.CPU26.CAL:Function_call_interrupts 6114 ± 5% -4.1% 5864 ± 5% interrupts.CPU27.CAL:Function_call_interrupts 6112 ± 5% -4.1% 5864 ± 5% interrupts.CPU28.CAL:Function_call_interrupts 6085 ± 6% -3.7% 5859 ± 5% interrupts.CPU29.CAL:Function_call_interrupts 780665 ± 20% -21.3% 614413 interrupts.CPU3.LOC:Local_timer_interrupts 1969 ± 90% -88.1% 235.00 ±106% interrupts.CPU3.RES:Rescheduling_interrupts 6085 ± 6% -3.8% 5855 ± 5% interrupts.CPU30.CAL:Function_call_interrupts 6079 ± 6% -3.7% 5853 ± 5% interrupts.CPU31.CAL:Function_call_interrupts 6080 ± 6% -3.8% 5847 ± 5% interrupts.CPU32.CAL:Function_call_interrupts 6077 ± 6% -3.7% 5850 ± 5% interrupts.CPU33.CAL:Function_call_interrupts 806.00 ± 11% -73.2% 216.00 ± 74% interrupts.CPU35.RES:Rescheduling_interrupts 1731 ± 3% -83.6% 283.75 ± 58% interrupts.CPU37.RES:Rescheduling_interrupts 45.50 ± 45% +1009.3% 504.75 ± 86% interrupts.CPU40.RES:Rescheduling_interrupts 41.50 ± 20% +222.3% 133.75 ± 47% interrupts.CPU44.RES:Rescheduling_interrupts 86.50 ± 26% +414.7% 445.25 ± 45% interrupts.CPU49.RES:Rescheduling_interrupts 129.50 ± 30% +1750.4% 2396 ± 94% interrupts.CPU53.RES:Rescheduling_interrupts 239.50 ± 51% +291.6% 938.00 ± 34% interrupts.CPU56.RES:Rescheduling_interrupts 6153 ± 5% -4.4% 5881 ± 5% interrupts.CPU58.CAL:Function_call_interrupts 87.00 ± 8% +1240.2% 1166 ± 58% interrupts.CPU63.RES:Rescheduling_interrupts 244.50 ± 29% -51.7% 118.00 ± 30% interrupts.CPU69.RES:Rescheduling_interrupts 1242 ± 64% -82.3% 220.00 ± 33% interrupts.CPU75.RES:Rescheduling_interrupts 173.50 ± 18% -43.4% 98.25 ± 10% interrupts.CPU76.RES:Rescheduling_interrupts 2162 ± 60% -76.4% 510.50 ±123% interrupts.CPU77.RES:Rescheduling_interrupts 74.00 ± 5% +285.5% 285.25 ± 45% interrupts.CPU8.RES:Rescheduling_interrupts 633.00 ± 34% -56.6% 274.50 ± 97% interrupts.CPU82.RES:Rescheduling_interrupts 6229 ± 7% -5.3% 5898 ± 5% interrupts.CPU84.CAL:Function_call_interrupts 6123 ± 5% -3.6% 5900 ± 5% interrupts.CPU86.CAL:Function_call_interrupts 943284 ± 34% -34.8% 614662 interrupts.CPU86.LOC:Local_timer_interrupts 6122 ± 5% -3.6% 5899 ± 5% interrupts.CPU87.CAL:Function_call_interrupts 6120 ± 5% -3.6% 5898 ± 5% interrupts.CPU88.CAL:Function_call_interrupts 6118 ± 5% -3.5% 5902 ± 5% interrupts.CPU89.CAL:Function_call_interrupts 1286 ± 83% -66.7% 428.50 ±118% interrupts.CPU9.RES:Rescheduling_interrupts 6114 ± 5% -3.5% 5900 ± 5% interrupts.CPU91.CAL:Function_call_interrupts 6118 ± 5% -3.4% 5908 ± 5% interrupts.CPU92.CAL:Function_call_interrupts 6117 ± 5% -3.4% 5907 ± 5% interrupts.CPU93.CAL:Function_call_interrupts 6118 ± 5% -3.4% 5907 ± 5% interrupts.CPU94.CAL:Function_call_interrupts 6074 ± 5% -3.8% 5841 ± 5% interrupts.CPU95.CAL:Function_call_interrupts 8980 ± 11% -9.6% 8117 ± 4% interrupts.CPU95.RES:Rescheduling_interrupts 6116 ± 5% -3.8% 5882 ± 5% interrupts.CPU96.CAL:Function_call_interrupts 6116 ± 5% -5.4% 5786 ± 2% interrupts.CPU97.CAL:Function_call_interrupts 6115 ± 5% -3.7% 5890 ± 5% interrupts.CPU99.CAL:Function_call_interrupts 846020 ± 26% -27.6% 612307 interrupts.CPU99.LOC:Local_timer_interrupts 311.00 -34.5% 203.75 ± 44% interrupts.TLB:TLB_shootdowns *************************************************************************************************** lkp-knm02: 272 threads Intel(R) Xeon Phi(TM) CPU 7255 @ 1.10GHz with 112G memory ========================================================================================= class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime: cpu/gcc-7/performance/1HDD/x86_64-rhel-7.2/100%/debian-x86_64-2018-04-03.cgz/lkp-knm02/stress-ng/1s commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- %stddev %change %stddev \ | \ 66.75 ± 8% +260.3% 240.50 ± 3% stress-ng.numa.ops 46.23 ± 5% +401.7% 231.92 ± 3% stress-ng.numa.ops_per_sec 44912 ± 3% +6.8% 47988 stress-ng.time.voluntary_context_switches 34176 ± 8% -38.7% 20951 ± 11% numa-numastat.node1.numa_hit 34176 ± 8% -38.7% 20951 ± 11% numa-numastat.node1.other_node 162.25 ± 13% -97.5% 4.00 ±106% numa-vmstat.node0.nr_isolated_anon 180.50 ± 13% -100.0% 0.00 numa-vmstat.node1.nr_isolated_anon 22.10 ± 5% -21.7% 17.31 ± 4% perf-stat.i.MPKI 14.96 -4.5% 14.28 perf-stat.overall.MPKI 7.58 +0.1 7.65 perf-stat.overall.branch-miss-rate% 18.33 +0.4 18.73 perf-stat.overall.cache-miss-rate% 1831 +2.2% 1871 perf-stat.overall.cycles-between-cache-misses 1.46 ± 33% -1.1 0.32 ±103% perf-profile.calltrace.cycles-pp.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock.vprintk_emit 1.44 ± 35% -0.8 0.63 ± 78% perf-profile.calltrace.cycles-pp.wait_for_xmitr.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock 1.70 ± 15% -0.6 1.11 ± 62% perf-profile.calltrace.cycles-pp.run_rebalance_domains.__softirqentry_text_start.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt 0.55 ± 66% +2.6 3.12 ± 75% perf-profile.calltrace.cycles-pp.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.__ioctl.perf_evlist__enable 0.55 ± 66% +2.6 3.13 ± 74% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__ioctl.perf_evlist__enable.cmd_record.run_builtin 0.55 ± 66% +2.6 3.13 ± 74% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__ioctl.perf_evlist__enable.cmd_record 2626 ± 10% +17.0% 3072 ± 6% slabinfo.avtab_node.active_objs 2626 ± 10% +17.0% 3072 ± 6% slabinfo.avtab_node.num_objs 44473 +7.0% 47586 ± 4% slabinfo.filp.active_objs 44478 +7.1% 47622 ± 4% slabinfo.filp.num_objs 828.75 ± 9% -14.2% 711.00 ± 4% slabinfo.skbuff_fclone_cache.active_objs 828.75 ± 9% -14.2% 711.00 ± 4% slabinfo.skbuff_fclone_cache.num_objs 431.00 ± 30% -99.0% 4.50 ±101% proc-vmstat.nr_isolated_anon 34176 ± 8% -38.7% 20951 ± 11% proc-vmstat.numa_other 23394485 ± 9% -74.5% 5964109 ± 40% proc-vmstat.pgalloc_normal 23265642 ± 9% -75.0% 5825344 ± 41% proc-vmstat.pgfree 592.75 ± 15% +472.5% 3393 ± 8% proc-vmstat.pgmigrate_fail 51229 ± 8% -86.2% 7061 ± 15% proc-vmstat.pgmigrate_success 40681 ± 10% -83.7% 6633 ± 71% proc-vmstat.thp_deferred_split_page 497.44 ± 3% -16.4% 415.64 ± 4% sched_debug.cfs_rq:/.exec_clock.stddev 4.59 ± 20% -24.7% 3.45 ± 10% sched_debug.cfs_rq:/.load_avg.avg 321.12 ± 46% -63.6% 116.75 ± 20% sched_debug.cfs_rq:/.load_avg.max 23.22 ± 35% -51.5% 11.26 ± 12% sched_debug.cfs_rq:/.load_avg.stddev 14.50 ± 22% +62.9% 23.62 ± 23% sched_debug.cfs_rq:/.nr_spread_over.max 1.68 ± 12% +26.8% 2.13 ± 18% sched_debug.cfs_rq:/.nr_spread_over.stddev 1078 ± 5% +17.5% 1267 ± 9% sched_debug.cfs_rq:/.util_avg.max 9496 ± 18% -15.6% 8019 ± 4% softirqs.CPU0.RCU 40692 ± 5% -9.7% 36753 ± 2% softirqs.CPU0.TIMER 40960 ± 5% -9.7% 36968 ± 5% softirqs.CPU10.TIMER 42943 ± 5% -7.9% 39563 ± 5% softirqs.CPU114.TIMER 7967 ± 15% -14.7% 6795 ± 5% softirqs.CPU128.RCU 44489 ± 5% -13.2% 38635 ± 5% softirqs.CPU128.TIMER 43555 ± 7% -13.6% 37648 ± 5% softirqs.CPU144.TIMER 7955 ± 16% -15.9% 6693 ± 4% softirqs.CPU145.RCU 43111 ± 7% -11.9% 37982 ± 4% softirqs.CPU160.TIMER 7755 ± 18% -14.0% 6667 ± 6% softirqs.CPU186.RCU 43744 ± 6% -7.9% 40296 ± 5% softirqs.CPU187.TIMER 41152 ± 7% -9.2% 37362 ± 4% softirqs.CPU190.TIMER 43610 ± 3% -11.9% 38408 ± 5% softirqs.CPU198.TIMER 42378 ± 7% -11.4% 37538 ± 4% softirqs.CPU240.TIMER 8735 ± 18% -21.6% 6845 ± 8% softirqs.CPU262.RCU 37455 ± 4% -7.7% 34575 ± 3% softirqs.CPU270.TIMER 8857 ± 20% -25.2% 6622 ± 5% softirqs.CPU271.RCU 40008 ± 5% -9.7% 36122 ± 2% softirqs.CPU32.TIMER 45805 ± 22% -17.4% 37825 ± 4% softirqs.CPU37.TIMER 40227 ± 4% -10.2% 36125 ± 4% softirqs.CPU55.TIMER 44908 ± 6% -12.3% 39385 ± 5% softirqs.CPU64.TIMER 12303 ± 8% -24.0% 9348 ± 18% softirqs.CPU69.RCU 51650 ± 4% -17.8% 42473 ± 2% softirqs.CPU69.TIMER 44028 ± 6% -9.6% 39794 ± 3% softirqs.CPU86.TIMER 2009 ± 12% -24.9% 1509 ± 17% interrupts.CPU0.RES:Rescheduling_interrupts 956.00 ±136% -91.7% 79.00 ± 24% interrupts.CPU10.RES:Rescheduling_interrupts 58.50 ± 38% +127.4% 133.00 ± 42% interrupts.CPU101.RES:Rescheduling_interrupts 33.25 ± 29% +226.3% 108.50 ± 94% interrupts.CPU104.RES:Rescheduling_interrupts 7122 ± 3% +10.9% 7897 ± 2% interrupts.CPU115.CAL:Function_call_interrupts 56.25 ± 27% +98.2% 111.50 ± 62% interrupts.CPU118.RES:Rescheduling_interrupts 446.25 ± 38% -48.8% 228.50 ± 18% interrupts.CPU118.TLB:TLB_shootdowns 7016 ± 5% +11.8% 7841 interrupts.CPU124.CAL:Function_call_interrupts 241.00 ±120% -77.3% 54.75 ± 23% interrupts.CPU124.RES:Rescheduling_interrupts 103.00 ± 41% -54.6% 46.75 ± 28% interrupts.CPU125.RES:Rescheduling_interrupts 37.00 ± 8% +303.4% 149.25 ± 76% interrupts.CPU156.RES:Rescheduling_interrupts 66.25 ± 81% +108.3% 138.00 ± 64% interrupts.CPU162.RES:Rescheduling_interrupts 34.75 ± 25% +656.1% 262.75 ±116% interrupts.CPU168.RES:Rescheduling_interrupts 117.25 ± 20% -42.9% 67.00 ± 29% interrupts.CPU193.RES:Rescheduling_interrupts 8393 ± 3% +11.5% 9359 ± 4% interrupts.CPU205.CAL:Function_call_interrupts 87.75 ± 39% -44.7% 48.50 ± 28% interrupts.CPU205.RES:Rescheduling_interrupts 6863 ± 5% +12.5% 7720 ± 2% interrupts.CPU21.CAL:Function_call_interrupts 47.25 ± 9% +66.7% 78.75 ± 27% interrupts.CPU224.RES:Rescheduling_interrupts 47.50 ± 33% +81.1% 86.00 ± 29% interrupts.CPU232.RES:Rescheduling_interrupts 32.25 ± 14% +106.2% 66.50 ± 50% interrupts.CPU241.RES:Rescheduling_interrupts 328.25 ± 25% +167.6% 878.50 ± 60% interrupts.CPU248.TLB:TLB_shootdowns 36.75 ± 43% +105.4% 75.50 ± 40% interrupts.CPU256.RES:Rescheduling_interrupts 377.75 ± 20% +29.8% 490.25 ± 15% interrupts.CPU264.TLB:TLB_shootdowns 73.50 ± 45% +204.8% 224.00 ± 61% interrupts.CPU28.RES:Rescheduling_interrupts 79.25 ± 37% +76.3% 139.75 ± 35% interrupts.CPU32.RES:Rescheduling_interrupts 6892 ± 6% +14.9% 7919 ± 3% interrupts.CPU37.CAL:Function_call_interrupts 6892 ± 4% +16.9% 8055 ± 8% interrupts.CPU39.CAL:Function_call_interrupts 51.75 ± 15% +68.1% 87.00 ± 34% interrupts.CPU39.RES:Rescheduling_interrupts 476.00 ± 17% +102.9% 966.00 ± 31% interrupts.CPU40.TLB:TLB_shootdowns 6847 ± 5% +12.6% 7711 ± 2% interrupts.CPU41.CAL:Function_call_interrupts 6853 ± 4% +10.9% 7603 ± 3% interrupts.CPU48.CAL:Function_call_interrupts 522.50 ± 27% -48.1% 271.00 ± 74% interrupts.CPU54.TLB:TLB_shootdowns 454.25 ± 29% +85.5% 842.75 ± 49% interrupts.CPU60.TLB:TLB_shootdowns 300.00 ± 32% +148.1% 744.25 ± 55% interrupts.CPU61.TLB:TLB_shootdowns 52.75 ± 23% +130.3% 121.50 ± 50% interrupts.CPU64.RES:Rescheduling_interrupts 52.00 ± 22% +122.6% 115.75 ± 62% interrupts.CPU70.RES:Rescheduling_interrupts 6972 ± 5% +8.9% 7595 ± 5% interrupts.CPU77.CAL:Function_call_interrupts 6923 ± 4% +10.5% 7650 ± 2% interrupts.CPU8.CAL:Function_call_interrupts 7233 ± 5% +11.3% 8049 ± 2% interrupts.CPU81.CAL:Function_call_interrupts 585.25 ± 10% -37.6% 365.00 ± 27% interrupts.CPU88.TLB:TLB_shootdowns *************************************************************************************************** lkp-knm02: 272 threads Intel(R) Xeon Phi(TM) CPU 7255 @ 1.10GHz with 112G memory ========================================================================================= class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime: pipe/gcc-7/performance/1HDD/x86_64-rhel-7.2/100%/debian-x86_64-2018-04-03.cgz/lkp-knm02/stress-ng/1s commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | :4 25% 1:4 kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#] :4 25% 1:4 dmesg.RIP:get_gup_pin_page 1:4 -25% :4 dmesg.WARNING:at#for_ip_swapgs_restore_regs_and_return_to_usermode/0x :4 25% 1:4 dmesg.WARNING:at_mm/gup.c:#get_gup_pin_page 1:4 -25% :4 dmesg.WARNING:stack_recursion %stddev %change %stddev \ | \ 18012 -10.5% 16117 ± 8% stress-ng.time.percent_of_cpu_this_job_got 1592 -4.9% 1513 stress-ng.time.system_time 3995345 -47.5% 2099227 stress-ng.vm-splice.ops 4001986 -18.5% 3259638 ± 7% stress-ng.vm-splice.ops_per_sec 0.24 ±173% +1.3 1.53 ± 44% perf-profile.calltrace.cycles-pp.cpu_load_update.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer 185055 ± 59% +38.4% 256144 ± 49% meminfo.Active 184967 ± 59% +38.4% 256057 ± 49% meminfo.Active(anon) 185946 ± 58% +40.8% 261876 ± 50% numa-meminfo.node0.Active 185858 ± 58% +40.9% 261789 ± 50% numa-meminfo.node0.Active(anon) 17098 ± 6% -11.5% 15128 ± 3% numa-meminfo.node1.SUnreclaim 47648 ± 62% +42.2% 67747 ± 51% numa-vmstat.node0.nr_active_anon 47622 ± 62% +42.2% 67725 ± 51% numa-vmstat.node0.nr_zone_active_anon 4274 ± 6% -11.5% 3782 ± 3% numa-vmstat.node1.nr_slab_unreclaimable 24880 +0.9% 25102 proc-vmstat.nr_slab_reclaimable 18690590 +1.3% 18925901 proc-vmstat.numa_hit 18690590 +1.3% 18925901 proc-vmstat.numa_local 10842 ± 3% +7.3% 11637 ± 5% softirqs.CPU13.TIMER 14451 ± 26% -26.5% 10615 ± 5% softirqs.CPU18.TIMER 10649 ± 5% +15.8% 12333 ± 10% softirqs.CPU56.TIMER 3.37 ± 7% -44.3% 1.88 ± 33% sched_debug.cfs_rq:/.exec_clock.avg 22.72 ± 6% -29.9% 15.92 ± 26% sched_debug.cfs_rq:/.exec_clock.stddev 5.80 ± 34% -40.9% 3.43 ± 11% sched_debug.cfs_rq:/.load_avg.avg 1020 -51.2% 498.00 ± 6% sched_debug.cfs_rq:/.util_est_enqueued.max 78.11 ± 12% -33.3% 52.06 ± 21% sched_debug.cfs_rq:/.util_est_enqueued.stddev 7243 ± 12% +79.0% 12964 ± 17% sched_debug.cpu.avg_idle.min 2420 ± 18% -19.7% 1943 perf-stat.i.cycles-between-cache-misses 137.18 ± 5% +16.2% 159.47 ± 6% perf-stat.i.instructions-per-iTLB-miss 1969 ± 4% -7.2% 1827 ± 2% perf-stat.overall.cycles-between-cache-misses 0.90 ± 3% -0.0 0.85 perf-stat.overall.iTLB-load-miss-rate% 111.89 ± 2% +4.8% 117.28 perf-stat.overall.instructions-per-iTLB-miss 1.822e+11 -12.8% 1.589e+11 ± 11% perf-stat.ps.cpu-cycles 10509 ± 3% -12.6% 9183 ± 10% perf-stat.ps.minor-faults 457963 ± 2% -12.0% 402874 ± 11% perf-stat.ps.msec 10647 ± 4% -13.0% 9267 ± 10% perf-stat.ps.page-faults 1840 ± 34% -35.6% 1184 ± 19% interrupts.CPU12.RES:Rescheduling_interrupts 792.50 ± 7% +33.7% 1059 ± 17% interrupts.CPU126.RES:Rescheduling_interrupts 743.00 ± 50% +119.2% 1628 ± 20% interrupts.CPU154.RES:Rescheduling_interrupts 1490 ± 9% -37.5% 931.50 ± 21% interrupts.CPU16.RES:Rescheduling_interrupts 880.50 ± 30% +70.6% 1502 ± 10% interrupts.CPU194.RES:Rescheduling_interrupts 931.00 ± 23% +49.5% 1391 ± 18% interrupts.CPU216.RES:Rescheduling_interrupts 1158 ± 11% -31.0% 798.75 ± 30% interrupts.CPU236.RES:Rescheduling_interrupts 1253 ± 21% -32.7% 843.00 ± 31% interrupts.CPU241.RES:Rescheduling_interrupts 1700 ± 37% -58.2% 710.00 ± 48% interrupts.CPU243.RES:Rescheduling_interrupts 1399 ± 20% -40.4% 834.25 ± 35% interrupts.CPU255.RES:Rescheduling_interrupts 1064 ± 19% +32.0% 1405 ± 7% interrupts.CPU256.RES:Rescheduling_interrupts 996.75 ± 24% +32.0% 1316 ± 11% interrupts.CPU262.RES:Rescheduling_interrupts 1129 ± 22% +46.2% 1652 ± 16% interrupts.CPU267.RES:Rescheduling_interrupts 1553 ± 19% -34.4% 1019 ± 25% interrupts.CPU39.RES:Rescheduling_interrupts 1110 ± 5% +25.3% 1390 ± 14% interrupts.CPU40.RES:Rescheduling_interrupts 1073 ± 18% +45.5% 1562 ± 4% interrupts.CPU41.RES:Rescheduling_interrupts 1478 ± 20% -35.3% 956.50 ± 22% interrupts.CPU47.RES:Rescheduling_interrupts 1554 ± 7% -29.6% 1094 ± 24% interrupts.CPU54.RES:Rescheduling_interrupts 1613 ± 18% -37.6% 1007 ± 38% interrupts.CPU57.RES:Rescheduling_interrupts 946.25 ± 15% +45.7% 1379 ± 10% interrupts.CPU64.RES:Rescheduling_interrupts 780.00 ± 31% +81.5% 1415 ± 12% interrupts.CPU84.RES:Rescheduling_interrupts *************************************************************************************************** lkp-bdw-ep3b: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory ========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase/ucode: gcc-7/performance/x86_64-rhel-7.2/process/100%/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3b/futex2/will-it-scale/0xb00002e commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- %stddev %change %stddev \ | \ 26674 ± 17% -22.3% 20716 ± 3% softirqs.CPU30.RCU 26755 ± 20% -21.3% 21064 ± 3% softirqs.CPU59.RCU 295164 +11.4% 328898 meminfo.Active 294996 +11.4% 328730 meminfo.Active(anon) 295368 -25.0% 221640 ± 14% meminfo.DirectMap4k 24.48 +0.2 24.66 perf-profile.calltrace.cycles-pp.get_futex_key.futex_wait_setup.futex_wait.do_futex.__x64_sys_futex 14.88 +0.2 15.09 perf-profile.calltrace.cycles-pp.gup_pgd_range.get_user_pages_fast.get_futex_key.futex_wait_setup.futex_wait 50.45 +0.2 50.67 perf-profile.calltrace.cycles-pp.do_futex.__x64_sys_futex.do_syscall_64.entry_SYSCALL_64_after_hwframe 73752 +11.4% 82177 proc-vmstat.nr_active_anon 73752 +11.4% 82177 proc-vmstat.nr_zone_active_anon 710645 -2.3% 694065 proc-vmstat.pgfree 2.108e+08 ± 15% -53.8% 97334654 ± 18% cpuidle.C6.time 262362 ± 3% -54.2% 120108 ± 14% cpuidle.C6.usage 27692 ± 77% -79.8% 5606 ± 27% cpuidle.POLL.time 4498 ± 40% -44.2% 2509 ± 8% cpuidle.POLL.usage 362291 -41.5% 211769 ± 18% numa-numastat.node0.local_node 379384 -42.7% 217528 ± 19% numa-numastat.node0.numa_hit 17093 -66.3% 5760 ±120% numa-numastat.node0.other_node 289553 +50.8% 436751 ± 9% numa-numastat.node1.local_node 289591 +54.7% 448130 ± 9% numa-numastat.node1.numa_hit 250685 ± 3% -56.5% 108986 ± 17% turbostat.C6 0.77 ± 16% -0.4 0.34 ± 20% turbostat.C6% 0.55 ± 24% -58.3% 0.23 ± 32% turbostat.CPU%c6 0.35 ± 23% -54.2% 0.16 ± 55% turbostat.Pkg%pc2 0.03 ± 66% -91.7% 0.00 ±173% turbostat.Pkg%pc6 34.58 ± 11% +30.4% 45.08 ± 4% sched_debug.cfs_rq:/.nr_spread_over.max 3.94 ± 7% +25.1% 4.92 ± 5% sched_debug.cfs_rq:/.nr_spread_over.stddev 2.39 ± 2% +183.0% 6.78 ±106% sched_debug.cpu.cpu_load[0].stddev 24.58 +107.3% 50.96 ± 62% sched_debug.cpu.cpu_load[1].max 2.26 ± 6% +118.1% 4.94 ± 64% sched_debug.cpu.cpu_load[1].stddev 25212 ± 22% -25.8% 18696 ± 27% sched_debug.cpu.sched_count.max 125453 -33.3% 83670 ± 19% numa-meminfo.node0.AnonHugePages 168582 -21.2% 132815 ± 14% numa-meminfo.node0.AnonPages 124539 +18.2% 147258 ± 4% numa-meminfo.node1.Active 124539 +18.2% 147216 ± 4% numa-meminfo.node1.Active(anon) 43851 ± 4% +99.6% 87509 ± 18% numa-meminfo.node1.AnonHugePages 75101 ± 2% +48.9% 111789 ± 16% numa-meminfo.node1.AnonPages 2816 ± 2% +22.4% 3447 ± 17% numa-meminfo.node1.PageTables 54950 +11.2% 61121 ± 7% numa-meminfo.node1.SUnreclaim 42139 -21.2% 33199 ± 14% numa-vmstat.node0.nr_anon_pages 711043 -19.2% 574713 ± 14% numa-vmstat.node0.numa_hit 693926 -18.0% 568865 ± 13% numa-vmstat.node0.numa_local 17117 -65.8% 5847 ±117% numa-vmstat.node0.numa_other 31111 +18.3% 36809 ± 4% numa-vmstat.node1.nr_active_anon 18777 ± 2% +48.8% 27945 ± 16% numa-vmstat.node1.nr_anon_pages 699.50 ± 3% +22.0% 853.50 ± 16% numa-vmstat.node1.nr_page_table_pages 13737 +11.2% 15279 ± 7% numa-vmstat.node1.nr_slab_unreclaimable 31111 +18.3% 36809 ± 4% numa-vmstat.node1.nr_zone_active_anon 426549 +32.8% 566367 ± 15% numa-vmstat.node1.numa_hit 288820 +44.7% 417867 ± 20% numa-vmstat.node1.numa_local 0.00 -0.0 0.00 ± 2% perf-stat.i.dTLB-load-miss-rate% 195436 -4.1% 187470 ± 3% perf-stat.i.dTLB-load-misses 67.36 -0.7 66.69 perf-stat.i.iTLB-load-miss-rate% 1.66e+08 +2.8% 1.707e+08 perf-stat.i.iTLB-loads 51056 -8.1% 46912 ± 2% perf-stat.i.node-load-misses 7153 -30.0% 5006 ± 21% perf-stat.i.node-loads 0.00 -0.0 0.00 ± 3% perf-stat.overall.dTLB-load-miss-rate% 67.37 -0.7 66.69 perf-stat.overall.iTLB-load-miss-rate% 194906 -4.1% 186955 ± 3% perf-stat.ps.dTLB-load-misses 1.655e+08 +2.8% 1.701e+08 perf-stat.ps.iTLB-loads 50904 -8.1% 46767 ± 2% perf-stat.ps.node-load-misses 7136 -30.0% 4996 ± 21% perf-stat.ps.node-loads 352.50 ± 10% +32.5% 467.00 ± 14% interrupts.32:PCI-MSI.3145729-edge.eth0-TxRx-0 222.00 ± 16% +84.0% 408.50 ± 27% interrupts.34:PCI-MSI.3145731-edge.eth0-TxRx-2 275.00 ± 12% -25.7% 204.25 ± 17% interrupts.37:PCI-MSI.3145734-edge.eth0-TxRx-5 352.50 ± 10% +32.5% 467.00 ± 14% interrupts.CPU11.32:PCI-MSI.3145729-edge.eth0-TxRx-0 651.50 ± 59% -76.4% 154.00 ± 59% interrupts.CPU11.RES:Rescheduling_interrupts 222.00 ± 16% +84.0% 408.50 ± 27% interrupts.CPU13.34:PCI-MSI.3145731-edge.eth0-TxRx-2 1011 ± 2% -82.7% 175.25 ± 50% interrupts.CPU13.RES:Rescheduling_interrupts 1680 ± 26% -64.1% 602.75 ± 85% interrupts.CPU14.RES:Rescheduling_interrupts 472.50 ± 53% -54.1% 217.00 ± 95% interrupts.CPU15.RES:Rescheduling_interrupts 275.00 ± 12% -25.7% 204.25 ± 17% interrupts.CPU16.37:PCI-MSI.3145734-edge.eth0-TxRx-5 635.00 ± 62% -54.5% 289.00 ±113% interrupts.CPU17.RES:Rescheduling_interrupts 903.00 ± 12% -60.4% 357.50 ± 82% interrupts.CPU18.RES:Rescheduling_interrupts 109.00 ± 30% +627.1% 792.50 ± 99% interrupts.CPU24.RES:Rescheduling_interrupts 146.00 ± 8% +365.9% 680.25 ± 62% interrupts.CPU28.RES:Rescheduling_interrupts 1609 ± 53% -79.9% 323.50 ± 83% interrupts.CPU3.RES:Rescheduling_interrupts 125.50 ± 25% +1227.3% 1665 ± 59% interrupts.CPU30.RES:Rescheduling_interrupts 381.00 ± 48% +240.0% 1295 ± 33% interrupts.CPU31.RES:Rescheduling_interrupts 127.50 ± 29% +1944.3% 2606 ± 69% interrupts.CPU36.RES:Rescheduling_interrupts 132.00 ± 22% +141.3% 318.50 ± 23% interrupts.CPU40.RES:Rescheduling_interrupts 255.50 ± 69% +377.2% 1219 ± 71% interrupts.CPU42.RES:Rescheduling_interrupts 965.00 ± 90% -93.8% 60.00 ± 93% interrupts.CPU45.RES:Rescheduling_interrupts 945.50 ± 96% -97.6% 22.75 ± 54% interrupts.CPU49.RES:Rescheduling_interrupts 231.50 ± 85% -88.1% 27.50 ± 89% interrupts.CPU50.RES:Rescheduling_interrupts 345.50 ± 91% -95.0% 17.25 ± 70% interrupts.CPU57.RES:Rescheduling_interrupts 71.50 ± 44% -47.6% 37.50 ± 88% interrupts.CPU59.RES:Rescheduling_interrupts 7877 -37.5% 4926 ± 34% interrupts.CPU6.NMI:Non-maskable_interrupts 7877 -37.5% 4926 ± 34% interrupts.CPU6.PMI:Performance_monitoring_interrupts 1167 ± 44% -88.0% 139.75 ± 60% interrupts.CPU6.RES:Rescheduling_interrupts 331.50 ± 91% -94.7% 17.50 ± 62% interrupts.CPU61.RES:Rescheduling_interrupts 136.50 ± 4% -81.7% 25.00 ± 57% interrupts.CPU64.RES:Rescheduling_interrupts 7897 -37.8% 4913 ± 34% interrupts.CPU7.NMI:Non-maskable_interrupts 7897 -37.8% 4913 ± 34% interrupts.CPU7.PMI:Performance_monitoring_interrupts 1356 ± 18% -65.8% 463.75 ± 6% interrupts.CPU7.RES:Rescheduling_interrupts 27.50 ± 5% +335.5% 119.75 ±101% interrupts.CPU73.RES:Rescheduling_interrupts 203.00 ± 13% -80.3% 40.00 ± 31% interrupts.CPU75.RES:Rescheduling_interrupts 7885 -37.4% 4934 ± 34% interrupts.CPU8.NMI:Non-maskable_interrupts 7885 -37.4% 4934 ± 34% interrupts.CPU8.PMI:Performance_monitoring_interrupts 1232 ± 51% -80.8% 236.50 ± 85% interrupts.CPU8.RES:Rescheduling_interrupts 295.00 ± 61% -58.7% 121.75 ± 65% interrupts.CPU87.RES:Rescheduling_interrupts *************************************************************************************************** lkp-knm02: 272 threads Intel(R) Xeon Phi(TM) CPU 7255 @ 1.10GHz with 112G memory ========================================================================================= class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime: os/gcc-7/performance/1HDD/x86_64-rhel-7.2/100%/debian-x86_64-2018-04-03.cgz/lkp-knm02/stress-ng/1s commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | 7:4 58% 10:4 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.key_lookup.lookup_user_key.keyctl_set_timeout :4 50% 2:4 dmesg.RIP:get_gup_pin_page :4 50% 2:4 dmesg.WARNING:at_mm/gup.c:#get_gup_pin_page :4 100% 4:4 kmsg.Memory_failure:#:dirty_LRU_page_still_referenced_by#users :4 100% 4:4 kmsg.Memory_failure:#:recovery_action_for_dirty_LRU_page:Failed 4:4 -100% :4 kmsg.Memory_failure:#:recovery_action_for_dirty_LRU_page:Recovered %stddev %change %stddev \ | \ 291197 -55.0% 131136 ± 11% stress-ng.futex.ops 290079 -58.0% 121851 ± 12% stress-ng.futex.ops_per_sec 2015 ± 80% +203.6% 6119 ± 61% stress-ng.io.ops_per_sec 122468 ± 2% +22.4% 149941 ± 2% stress-ng.mlock.ops_per_sec 65.00 ± 6% +251.2% 228.25 ± 5% stress-ng.numa.ops 45.61 ± 5% +383.3% 220.41 ± 5% stress-ng.numa.ops_per_sec 625254 ± 6% -23.1% 480881 ± 10% stress-ng.sem.ops 624395 ± 6% -23.1% 480000 ± 10% stress-ng.sem.ops_per_sec 14903791 -2.6% 14517111 stress-ng.time.voluntary_context_switches 306260 ± 4% -8.3% 280934 ± 5% stress-ng.userfaultfd.ops 306179 ± 4% -8.2% 280922 ± 5% stress-ng.userfaultfd.ops_per_sec 4157856 -49.5% 2098955 stress-ng.vm-splice.ops 4161453 -13.4% 3603810 stress-ng.vm-splice.ops_per_sec 1.53e+08 ± 2% +3.1% 1.578e+08 perf-stat.ps.iTLB-load-misses 4293169 ± 66% -44.9% 2363760 ± 2% cpuidle.POLL.time 81481 ± 48% -34.3% 53539 ± 2% cpuidle.POLL.usage 33283 ± 6% -59.2% 13575 ± 11% numa-numastat.node1.numa_hit 33283 ± 6% -59.2% 13575 ± 11% numa-numastat.node1.other_node 564127 ± 4% +185.9% 1612672 ± 19% meminfo.Active 563846 ± 4% +185.9% 1612232 ± 19% meminfo.Active(anon) 485.00 ± 22% +118.9% 1061 ± 7% meminfo.Inactive(file) 4144644 +26.3% 5236228 ± 6% meminfo.Memused 510156 ± 3% +205.4% 1558069 ± 20% numa-meminfo.node0.Active 509874 ± 3% +205.5% 1557628 ± 20% numa-meminfo.node0.Active(anon) 474.50 ± 21% +123.8% 1061 ± 9% numa-meminfo.node0.Inactive(file) 3534355 +30.6% 4617529 ± 7% numa-meminfo.node0.MemUsed 123321 ± 4% +218.3% 392482 ± 20% numa-vmstat.node0.nr_active_anon 70.00 ± 19% +58.9% 111.25 ± 3% numa-vmstat.node0.nr_active_file 119.00 ± 21% +124.2% 266.75 ± 8% numa-vmstat.node0.nr_inactive_file 123242 ± 4% +218.4% 392451 ± 20% numa-vmstat.node0.nr_zone_active_anon 70.00 ± 19% +58.9% 111.25 ± 3% numa-vmstat.node0.nr_zone_active_file 119.00 ± 21% +124.2% 266.75 ± 8% numa-vmstat.node0.nr_zone_inactive_file 3053 ± 6% -14.4% 2615 ± 12% sched_debug.cfs_rq:/.exec_clock.stddev 1.46 ± 6% -15.1% 1.24 ± 5% sched_debug.cfs_rq:/.nr_running.max 585.54 ± 15% -35.4% 378.27 ± 19% sched_debug.cfs_rq:/.runnable_load_avg.max 567.38 ± 15% -37.7% 353.29 ± 11% sched_debug.cpu.cpu_load[0].max 93234 ± 21% -24.0% 70869 ± 10% sched_debug.cpu.curr->pid.avg 185032 ± 4% -13.7% 159654 ± 10% sched_debug.cpu.curr->pid.max 23033 ± 9% -9.2% 20906 ± 6% sched_debug.cpu.sched_goidle.max 2420 ± 6% -13.9% 2085 ± 4% sched_debug.cpu.sched_goidle.stddev 38.68 ± 40% -60.9% 15.14 ± 30% sched_debug.rt_rq:/.rt_time.max 7.03 ± 67% -61.5% 2.71 ± 40% sched_debug.rt_rq:/.rt_time.stddev 134783 ± 5% +190.2% 391176 ± 17% proc-vmstat.nr_active_anon 70.25 ± 17% +58.4% 111.25 ± 4% proc-vmstat.nr_active_file 121.75 ± 20% +117.9% 265.25 ± 7% proc-vmstat.nr_inactive_file 131.00 ± 23% -99.8% 0.25 ±173% proc-vmstat.nr_isolated_anon 105352 -2.8% 102442 proc-vmstat.nr_slab_reclaimable 134783 ± 5% +190.2% 391176 ± 17% proc-vmstat.nr_zone_active_anon 70.25 ± 17% +58.4% 111.25 ± 4% proc-vmstat.nr_zone_active_file 121.75 ± 20% +117.9% 265.25 ± 7% proc-vmstat.nr_zone_inactive_file 864.50 ± 25% +638.5% 6384 ± 83% proc-vmstat.numa_huge_pte_updates 33283 ± 6% -59.2% 13575 ± 11% proc-vmstat.numa_other 445254 ± 25% +634.9% 3272027 ± 83% proc-vmstat.numa_pte_updates 466.00 ± 12% +382.9% 2250 ± 9% proc-vmstat.pgmigrate_fail 49887 ± 6% -92.6% 3698 ± 33% proc-vmstat.pgmigrate_success 7822768 -1.3% 7721955 proc-vmstat.unevictable_pgs_culled 352423 ± 2% +4.2% 367091 proc-vmstat.unevictable_pgs_mlocked 351801 ± 2% +4.1% 366356 proc-vmstat.unevictable_pgs_munlocked 351393 ± 2% +4.1% 365940 proc-vmstat.unevictable_pgs_rescued 11075 ± 8% -16.5% 9245 ± 9% softirqs.CPU100.NET_RX 8323 ± 9% +47.3% 12259 ± 23% softirqs.CPU111.NET_RX 10330 ± 11% -17.8% 8494 ± 5% softirqs.CPU120.NET_RX 28219 ± 4% -13.9% 24292 ± 3% softirqs.CPU123.RCU 19316 ± 6% -8.9% 17600 ± 4% softirqs.CPU127.SCHED 11518 ± 2% -26.4% 8477 ± 9% softirqs.CPU128.NET_RX 6724 ± 22% +67.4% 11258 ± 24% softirqs.CPU131.NET_RX 8571 ± 33% +62.2% 13903 ± 24% softirqs.CPU138.NET_RX 7441 ± 4% +33.9% 9965 ± 14% softirqs.CPU141.NET_RX 8270 ± 16% +44.7% 11966 ± 15% softirqs.CPU146.NET_RX 11298 ± 20% -36.7% 7155 ± 18% softirqs.CPU148.NET_RX 8935 ± 23% +33.2% 11900 ± 29% softirqs.CPU153.NET_RX 19787 ± 5% -9.6% 17897 ± 3% softirqs.CPU163.SCHED 26941 ± 5% -9.5% 24385 ± 3% softirqs.CPU169.RCU 12370 ± 25% -34.7% 8074 ± 13% softirqs.CPU177.NET_RX 20748 ± 6% -10.9% 18487 ± 4% softirqs.CPU181.SCHED 20707 ± 9% -16.1% 17369 ± 2% softirqs.CPU186.SCHED 19622 ± 6% -10.7% 17522 ± 3% softirqs.CPU187.SCHED 19406 ± 6% -8.9% 17672 ± 2% softirqs.CPU194.SCHED 32272 ± 8% -14.2% 27679 ± 7% softirqs.CPU20.RCU 12603 ± 21% -29.4% 8893 ± 16% softirqs.CPU209.NET_RX 11432 ± 9% -14.7% 9749 ± 14% softirqs.CPU21.NET_RX 9075 ± 9% -19.4% 7310 ± 7% softirqs.CPU214.NET_RX 8118 ± 9% +55.0% 12584 ± 24% softirqs.CPU240.NET_RX 9232 ± 16% -27.3% 6708 ± 9% softirqs.CPU244.NET_RX 8203 ± 15% +32.8% 10895 ± 7% softirqs.CPU265.NET_RX 14174 ± 8% -31.4% 9727 ± 23% softirqs.CPU267.NET_RX 17401 ± 11% -55.9% 7675 ± 21% softirqs.CPU28.NET_RX 30370 ± 4% -12.7% 26507 ± 3% softirqs.CPU3.RCU 11791 ± 4% -34.4% 7732 ± 13% softirqs.CPU46.NET_RX 19641 ± 7% -10.5% 17574 ± 4% softirqs.CPU51.SCHED 9879 ± 18% +21.5% 12003 ± 13% softirqs.CPU53.NET_RX 9572 ± 7% -8.3% 8781 ± 6% softirqs.CPU58.NET_RX 11786 ± 12% -33.7% 7818 ± 4% softirqs.CPU6.NET_RX 19919 ± 8% -12.9% 17341 ± 4% softirqs.CPU69.SCHED 8365 ± 15% +64.2% 13734 ± 35% softirqs.CPU76.NET_RX 7648 ± 11% +38.0% 10557 ± 17% softirqs.CPU77.NET_RX 9286 ± 15% +44.2% 13394 ± 16% softirqs.CPU88.NET_RX 27697 ± 4% -11.6% 24482 ± 3% softirqs.CPU90.RCU 8763 ± 16% +34.9% 11818 ± 22% softirqs.CPU94.NET_RX 2680 ± 15% -51.8% 1293 ± 28% interrupts.CPU100.NMI:Non-maskable_interrupts 2680 ± 15% -51.8% 1293 ± 28% interrupts.CPU100.PMI:Performance_monitoring_interrupts 2219 ± 20% -30.4% 1545 ± 22% interrupts.CPU104.NMI:Non-maskable_interrupts 2219 ± 20% -30.4% 1545 ± 22% interrupts.CPU104.PMI:Performance_monitoring_interrupts 2314 ± 19% -31.7% 1580 ± 10% interrupts.CPU106.TLB:TLB_shootdowns 2699 ± 18% -30.8% 1868 ± 10% interrupts.CPU108.TLB:TLB_shootdowns 17763 ± 25% +63.0% 28961 ± 13% interrupts.CPU113.RES:Rescheduling_interrupts 16871 ± 22% +88.1% 31734 ± 27% interrupts.CPU125.RES:Rescheduling_interrupts 24894 ± 18% +27.6% 31769 ± 8% interrupts.CPU128.RES:Rescheduling_interrupts 1755 ± 23% +77.4% 3112 ± 57% interrupts.CPU130.TLB:TLB_shootdowns 2644 ± 31% -35.0% 1719 ± 20% interrupts.CPU142.TLB:TLB_shootdowns 1164 ± 14% +64.3% 1912 ± 46% interrupts.CPU145.NMI:Non-maskable_interrupts 1164 ± 14% +64.3% 1912 ± 46% interrupts.CPU145.PMI:Performance_monitoring_interrupts 23656 ± 38% +53.7% 36365 ± 30% interrupts.CPU150.RES:Rescheduling_interrupts 27724 ± 6% -30.2% 19354 ± 16% interrupts.CPU151.RES:Rescheduling_interrupts 2491 ± 23% -29.6% 1754 ± 18% interrupts.CPU155.TLB:TLB_shootdowns 2153 ± 17% +16.1% 2499 ± 16% interrupts.CPU167.TLB:TLB_shootdowns 2170 ± 18% -20.0% 1735 ± 11% interrupts.CPU169.TLB:TLB_shootdowns 2004 ± 30% -33.3% 1337 ± 17% interrupts.CPU170.NMI:Non-maskable_interrupts 2004 ± 30% -33.3% 1337 ± 17% interrupts.CPU170.PMI:Performance_monitoring_interrupts 2079 ± 20% -25.1% 1557 ± 8% interrupts.CPU171.TLB:TLB_shootdowns 1754 ± 10% +44.1% 2527 ± 17% interrupts.CPU187.TLB:TLB_shootdowns 1632 ± 12% +45.6% 2376 ± 21% interrupts.CPU191.TLB:TLB_shootdowns 3271 ± 49% -50.5% 1618 ± 49% interrupts.CPU199.NMI:Non-maskable_interrupts 3271 ± 49% -50.5% 1618 ± 49% interrupts.CPU199.PMI:Performance_monitoring_interrupts 3481 ± 41% -52.1% 1668 ± 33% interrupts.CPU200.NMI:Non-maskable_interrupts 3481 ± 41% -52.1% 1668 ± 33% interrupts.CPU200.PMI:Performance_monitoring_interrupts 20086 ± 22% +84.1% 36982 ± 5% interrupts.CPU212.RES:Rescheduling_interrupts 1738 ± 12% +17.8% 2048 ± 5% interrupts.CPU212.TLB:TLB_shootdowns 18973 ± 23% +51.8% 28802 ± 19% interrupts.CPU216.RES:Rescheduling_interrupts 2301 ± 28% -51.6% 1113 ± 41% interrupts.CPU218.NMI:Non-maskable_interrupts 2301 ± 28% -51.6% 1113 ± 41% interrupts.CPU218.PMI:Performance_monitoring_interrupts 2267 ± 10% -16.8% 1887 ± 15% interrupts.CPU220.TLB:TLB_shootdowns 15399 ± 14% +61.3% 24836 ± 7% interrupts.CPU222.RES:Rescheduling_interrupts 36076 ± 16% -43.8% 20263 ± 36% interrupts.CPU226.RES:Rescheduling_interrupts 19194 ± 10% +84.8% 35478 ± 13% interrupts.CPU227.RES:Rescheduling_interrupts 2425 ± 8% -19.9% 1942 ± 6% interrupts.CPU227.TLB:TLB_shootdowns 2303 ± 22% -23.9% 1753 ± 8% interrupts.CPU234.TLB:TLB_shootdowns 2837 ± 8% -49.9% 1421 ± 11% interrupts.CPU239.NMI:Non-maskable_interrupts 2837 ± 8% -49.9% 1421 ± 11% interrupts.CPU239.PMI:Performance_monitoring_interrupts 17386 ± 17% +84.6% 32093 ± 30% interrupts.CPU240.RES:Rescheduling_interrupts 18801 ± 55% +92.8% 36248 ± 23% interrupts.CPU246.RES:Rescheduling_interrupts 2469 ± 31% -29.8% 1732 ± 17% interrupts.CPU247.TLB:TLB_shootdowns 31757 ± 27% -35.3% 20541 ± 31% interrupts.CPU250.RES:Rescheduling_interrupts 36905 ± 40% -46.7% 19670 ± 8% interrupts.CPU252.RES:Rescheduling_interrupts 20156 ± 12% +44.7% 29156 ± 7% interrupts.CPU253.RES:Rescheduling_interrupts 4880 ± 30% -48.0% 2539 ± 10% interrupts.CPU26.TLB:TLB_shootdowns 30073 ± 26% -37.2% 18889 ± 28% interrupts.CPU260.RES:Rescheduling_interrupts 24487 ± 4% -23.7% 18693 ± 23% interrupts.CPU261.RES:Rescheduling_interrupts 2201 ± 26% -42.9% 1257 ± 29% interrupts.CPU262.NMI:Non-maskable_interrupts 2201 ± 26% -42.9% 1257 ± 29% interrupts.CPU262.PMI:Performance_monitoring_interrupts 16758 ± 30% +76.3% 29544 ± 16% interrupts.CPU263.RES:Rescheduling_interrupts 2958 ± 56% -54.4% 1350 ± 25% interrupts.CPU266.NMI:Non-maskable_interrupts 2958 ± 56% -54.4% 1350 ± 25% interrupts.CPU266.PMI:Performance_monitoring_interrupts 29594 ± 23% -35.9% 18979 ± 33% interrupts.CPU268.RES:Rescheduling_interrupts 2391 ± 20% -42.8% 1367 ± 25% interrupts.CPU27.NMI:Non-maskable_interrupts 2391 ± 20% -42.8% 1367 ± 25% interrupts.CPU27.PMI:Performance_monitoring_interrupts 17135 ± 26% +74.3% 29873 ± 20% interrupts.CPU32.RES:Rescheduling_interrupts 32963 ± 29% -31.4% 22602 ± 27% interrupts.CPU36.RES:Rescheduling_interrupts 2328 ± 28% -34.4% 1527 ± 50% interrupts.CPU39.NMI:Non-maskable_interrupts 2328 ± 28% -34.4% 1527 ± 50% interrupts.CPU39.PMI:Performance_monitoring_interrupts 16869 ± 24% +69.0% 28507 ± 22% interrupts.CPU39.RES:Rescheduling_interrupts 2584 ± 21% -26.7% 1895 ± 16% interrupts.CPU51.TLB:TLB_shootdowns 3155 ± 16% -36.1% 2015 ± 20% interrupts.CPU52.TLB:TLB_shootdowns 4951 ± 28% -60.4% 1961 ± 25% interrupts.CPU55.TLB:TLB_shootdowns 4382 ± 24% -50.2% 2183 ± 14% interrupts.CPU56.TLB:TLB_shootdowns 15835 ± 21% +108.6% 33033 ± 30% interrupts.CPU6.RES:Rescheduling_interrupts 3549 ± 29% -34.9% 2310 ± 16% interrupts.CPU61.TLB:TLB_shootdowns 3466 ± 41% -51.2% 1691 ± 58% interrupts.CPU67.NMI:Non-maskable_interrupts 3466 ± 41% -51.2% 1691 ± 58% interrupts.CPU67.PMI:Performance_monitoring_interrupts 35989 ± 19% -36.0% 23024 ± 27% interrupts.CPU75.RES:Rescheduling_interrupts 2210 ± 6% -30.4% 1538 ± 14% interrupts.CPU76.TLB:TLB_shootdowns 3307 ± 45% -62.2% 1250 ± 14% interrupts.CPU78.NMI:Non-maskable_interrupts 3307 ± 45% -62.2% 1250 ± 14% interrupts.CPU78.PMI:Performance_monitoring_interrupts 1604 +54.6% 2480 ± 12% interrupts.CPU78.TLB:TLB_shootdowns 2863 ± 57% -42.8% 1636 ± 11% interrupts.CPU90.NMI:Non-maskable_interrupts 2863 ± 57% -42.8% 1636 ± 11% interrupts.CPU90.PMI:Performance_monitoring_interrupts 2308 ± 11% -21.7% 1808 ± 8% interrupts.CPU91.TLB:TLB_shootdowns 18040 ± 36% +89.6% 34197 ± 9% interrupts.CPU94.RES:Rescheduling_interrupts 21085 ± 18% +51.0% 31835 ± 16% interrupts.CPU97.RES:Rescheduling_interrupts *************************************************************************************************** lkp-bdw-ep3: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory ========================================================================================= class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode: vm/gcc-7/performance/1HDD/x86_64-rhel-7.2/100%/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3/stress-ng/1s/0xb00002e commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | :4 100% 4:4 kmsg.Memory_failure:#:dirty_LRU_page_still_referenced_by#users :4 100% 4:4 kmsg.Memory_failure:#:recovery_action_for_dirty_LRU_page:Failed 4:4 -100% :4 kmsg.Memory_failure:#:recovery_action_for_dirty_LRU_page:Recovered :4 100% 4:4 dmesg.RIP:get_gup_pin_page 1:4 -25% :4 dmesg.WARNING:at#for_ip_swapgs_restore_regs_and_return_to_usermode/0x :4 100% 4:4 dmesg.WARNING:at_mm/gup.c:#get_gup_pin_page 1:4 -25% :4 dmesg.WARNING:stack_recursion %stddev %change %stddev \ | \ 7765 -3.4% 7499 stress-ng.time.percent_of_cpu_this_job_got 1880 -5.1% 1784 stress-ng.time.system_time 5023675 ± 2% -58.3% 2095111 stress-ng.vm-splice.ops 17.91 ± 8% +12.9% 20.22 ± 6% iostat.cpu.idle 2406 -3.4% 2325 turbostat.Avg_MHz 59555797 -11.7% 52587698 vmstat.memory.free 15.47 ± 9% +2.3 17.80 ± 7% mpstat.cpu.idle% 0.10 ± 8% -0.0 0.08 ± 8% mpstat.cpu.soft% 34300 ± 2% -5.0% 32577 perf-stat.ps.major-faults 170288 -0.9% 168728 perf-stat.ps.msec 1131 ± 7% -17.7% 930.75 ± 14% slabinfo.nsproxy.active_objs 1131 ± 7% -17.7% 930.75 ± 14% slabinfo.nsproxy.num_objs 4662 ± 9% +69.2% 7887 ± 49% softirqs.CPU22.RCU 3963 ± 8% +151.7% 9974 ± 50% softirqs.CPU78.RCU 2905249 ± 10% +192.0% 8483775 ± 3% meminfo.Active 2905081 ± 10% +192.0% 8483607 ± 3% meminfo.Active(anon) 23150145 ± 4% -9.9% 20851826 ± 3% meminfo.Committed_AS 1365340 ± 5% +84.5% 2518784 ± 4% meminfo.Inactive 1364952 ± 5% +84.1% 2512920 ± 4% meminfo.Inactive(anon) 387.25 ± 6% +1414.2% 5863 ± 4% meminfo.Inactive(file) 59176014 -11.4% 52417164 meminfo.MemAvailable 59432778 -11.4% 52671931 meminfo.MemFree 6419352 ± 3% +105.3% 13180195 ± 2% meminfo.Memused 1.88 ± 6% +110.3% 3.96 ± 39% sched_debug.cpu.cpu_load[1].avg 19.25 ± 2% +629.9% 140.50 ± 62% sched_debug.cpu.cpu_load[1].max 4.41 ± 4% +258.0% 15.78 ± 53% sched_debug.cpu.cpu_load[1].stddev 2.31 ± 14% +94.2% 4.48 ± 31% sched_debug.cpu.cpu_load[2].avg 26.25 ± 47% +471.4% 150.00 ± 41% sched_debug.cpu.cpu_load[2].max 4.71 ± 21% +248.5% 16.42 ± 38% sched_debug.cpu.cpu_load[2].stddev 37.25 ± 37% +208.1% 114.75 ± 37% sched_debug.cpu.cpu_load[3].max 5.87 ± 19% +125.8% 13.25 ± 34% sched_debug.cpu.cpu_load[3].stddev 33.00 ± 19% +122.0% 73.25 ± 46% sched_debug.cpu.cpu_load[4].max 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.static_key_slow_dec.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.__static_key_slow_dec_cpuslocked.static_key_slow_dec.sw_perf_event_destroy._free_event.perf_event_release_kernel 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.__jump_label_update.__static_key_slow_dec_cpuslocked.static_key_slow_dec.sw_perf_event_destroy._free_event 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.arch_jump_label_transform.__jump_label_update.__static_key_slow_dec_cpuslocked.static_key_slow_dec.sw_perf_event_destroy 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.__jump_label_transform.arch_jump_label_transform.__jump_label_update.__static_key_slow_dec_cpuslocked.static_key_slow_dec 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.smp_call_function_many.on_each_cpu.text_poke_bp.__jump_label_transform.arch_jump_label_transform 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.on_each_cpu.text_poke_bp.__jump_label_transform.arch_jump_label_transform.__jump_label_update 11.46 ± 47% -7.5 3.97 ± 98% perf-profile.calltrace.cycles-pp.text_poke_bp.__jump_label_transform.arch_jump_label_transform.__jump_label_update.__static_key_slow_dec_cpuslocked 11.46 ± 47% -6.7 4.78 ± 69% perf-profile.calltrace.cycles-pp._free_event.perf_event_release_kernel.perf_release.__fput.task_work_run 11.46 ± 47% -6.7 4.78 ± 69% perf-profile.calltrace.cycles-pp.sw_perf_event_destroy._free_event.perf_event_release_kernel.perf_release.__fput 695203 ± 5% +213.3% 2177910 ± 3% proc-vmstat.nr_active_anon 1473798 -11.8% 1299346 proc-vmstat.nr_dirty_background_threshold 2952003 -11.8% 2603072 proc-vmstat.nr_dirty_threshold 14852961 -11.8% 13107633 proc-vmstat.nr_free_pages 377237 ± 7% +69.4% 639157 ± 3% proc-vmstat.nr_inactive_anon 97.00 ± 3% +1395.6% 1450 ± 6% proc-vmstat.nr_inactive_file 695202 ± 5% +213.3% 2177910 ± 3% proc-vmstat.nr_zone_active_anon 377237 ± 7% +69.4% 639157 ± 3% proc-vmstat.nr_zone_inactive_anon 97.00 ± 3% +1395.6% 1450 ± 6% proc-vmstat.nr_zone_inactive_file 12792774 ± 2% -10.7% 11419593 ± 7% proc-vmstat.pgactivate 2113 -3.3% 2043 ± 2% proc-vmstat.thp_split_page 1404817 ± 4% +213.6% 4405232 ± 9% numa-meminfo.node0.Active 1404733 ± 4% +213.6% 4405148 ± 9% numa-meminfo.node0.Active(anon) 828406 ± 12% +51.0% 1250665 ± 10% numa-meminfo.node0.Inactive 828190 ± 12% +50.7% 1247741 ± 10% numa-meminfo.node0.Inactive(anon) 214.75 ± 59% +1261.6% 2924 ± 10% numa-meminfo.node0.Inactive(file) 29574615 -11.6% 26131731 ± 2% numa-meminfo.node0.MemFree 3284371 ± 6% +104.8% 6727255 ± 8% numa-meminfo.node0.MemUsed 1476404 ± 11% +204.2% 4491068 ± 2% numa-meminfo.node1.Active 1476320 ± 11% +204.2% 4490984 ± 2% numa-meminfo.node1.Active(anon) 684877 ± 19% +87.8% 1286317 ± 9% numa-meminfo.node1.Inactive 684690 ± 19% +87.5% 1283519 ± 9% numa-meminfo.node1.Inactive(anon) 186.50 ± 65% +1400.3% 2798 ± 6% numa-meminfo.node1.Inactive(file) 29743730 -12.1% 26154353 numa-meminfo.node1.MemFree 3249412 ± 9% +110.5% 6838786 ± 3% numa-meminfo.node1.MemUsed 377498 ± 12% +193.5% 1108039 ± 9% numa-vmstat.node0.nr_active_anon 7374428 -11.6% 6515488 ± 2% numa-vmstat.node0.nr_free_pages 203915 ± 20% +57.8% 321729 ± 8% numa-vmstat.node0.nr_inactive_anon 55.25 ± 59% +1211.3% 724.50 ± 11% numa-vmstat.node0.nr_inactive_file 377455 ± 12% +193.5% 1108012 ± 9% numa-vmstat.node0.nr_zone_active_anon 203887 ± 20% +57.8% 321697 ± 8% numa-vmstat.node0.nr_zone_inactive_anon 55.25 ± 59% +1211.3% 724.50 ± 11% numa-vmstat.node0.nr_zone_inactive_file 392594 ± 12% +181.8% 1106200 ± 3% numa-vmstat.node1.nr_active_anon 7400171 -11.6% 6543657 numa-vmstat.node1.nr_free_pages 185991 ± 20% +78.6% 332154 ± 8% numa-vmstat.node1.nr_inactive_anon 44.75 ± 65% +1458.7% 697.50 ± 5% numa-vmstat.node1.nr_inactive_file 392558 ± 12% +181.8% 1106188 ± 3% numa-vmstat.node1.nr_zone_active_anon 185980 ± 20% +78.6% 332130 ± 8% numa-vmstat.node1.nr_zone_inactive_anon 44.75 ± 65% +1458.7% 697.50 ± 5% numa-vmstat.node1.nr_zone_inactive_file 92917 -2.7% 90421 interrupts.CAL:Function_call_interrupts 537.00 ± 10% +28.6% 690.50 ± 12% interrupts.CPU13.RES:Rescheduling_interrupts 574.25 ± 6% +44.4% 829.50 ± 29% interrupts.CPU17.RES:Rescheduling_interrupts 4162 ± 12% -22.4% 3231 ± 5% interrupts.CPU22.TLB:TLB_shootdowns 4051 ± 11% -20.9% 3203 ± 3% interrupts.CPU23.TLB:TLB_shootdowns 4017 ± 5% -17.4% 3318 ± 3% interrupts.CPU24.TLB:TLB_shootdowns 4249 ± 10% -22.7% 3283 ± 8% interrupts.CPU25.TLB:TLB_shootdowns 4003 ± 7% -17.8% 3291 ± 2% interrupts.CPU26.TLB:TLB_shootdowns 4161 ± 8% -19.8% 3336 ± 7% interrupts.CPU27.TLB:TLB_shootdowns 4063 ± 5% -17.9% 3337 ± 5% interrupts.CPU28.TLB:TLB_shootdowns 3970 ± 6% -17.6% 3271 ± 4% interrupts.CPU29.TLB:TLB_shootdowns 4093 ± 7% -20.1% 3270 ± 3% interrupts.CPU30.TLB:TLB_shootdowns 4001 ± 7% -15.2% 3392 ± 9% interrupts.CPU31.TLB:TLB_shootdowns 4246 ± 7% -22.6% 3284 ± 3% interrupts.CPU32.TLB:TLB_shootdowns 3949 ± 6% -14.9% 3361 ± 6% interrupts.CPU33.TLB:TLB_shootdowns 3949 ± 5% -15.6% 3332 ± 8% interrupts.CPU34.TLB:TLB_shootdowns 4057 ± 9% -20.4% 3228 ± 3% interrupts.CPU35.TLB:TLB_shootdowns 4105 ± 7% -22.5% 3180 ± 8% interrupts.CPU36.TLB:TLB_shootdowns 3957 ± 7% -18.7% 3217 ± 7% interrupts.CPU37.TLB:TLB_shootdowns 4071 ± 8% -20.7% 3229 ± 4% interrupts.CPU38.TLB:TLB_shootdowns 4256 ± 18% -22.8% 3286 ± 4% interrupts.CPU39.TLB:TLB_shootdowns 4334 ± 2% -22.0% 3381 ± 6% interrupts.CPU40.TLB:TLB_shootdowns 3858 ± 5% -17.2% 3195 ± 7% interrupts.CPU42.TLB:TLB_shootdowns 4212 ± 9% -20.3% 3356 ± 3% interrupts.CPU43.TLB:TLB_shootdowns 3922 ± 9% -17.7% 3227 ± 4% interrupts.CPU66.TLB:TLB_shootdowns 4021 ± 3% -16.1% 3372 ± 5% interrupts.CPU67.TLB:TLB_shootdowns 4176 ± 4% -19.0% 3381 ± 4% interrupts.CPU68.TLB:TLB_shootdowns 4149 ± 5% -39.5% 2508 ± 57% interrupts.CPU69.TLB:TLB_shootdowns 3993 ± 8% -18.1% 3270 ± 3% interrupts.CPU70.TLB:TLB_shootdowns 4059 ± 4% -17.4% 3352 ± 4% interrupts.CPU71.TLB:TLB_shootdowns 514.50 ± 5% +49.7% 770.00 ± 24% interrupts.CPU72.RES:Rescheduling_interrupts 3994 ± 5% -20.9% 3158 ± 2% interrupts.CPU72.TLB:TLB_shootdowns 4034 ± 7% -17.1% 3342 ± 3% interrupts.CPU73.TLB:TLB_shootdowns 4018 ± 7% -15.1% 3412 ± 4% interrupts.CPU74.TLB:TLB_shootdowns 4052 ± 7% -14.6% 3459 ± 5% interrupts.CPU75.TLB:TLB_shootdowns 4060 ± 6% -19.6% 3263 ± 5% interrupts.CPU76.TLB:TLB_shootdowns 4025 ± 5% -20.1% 3217 ± 2% interrupts.CPU77.TLB:TLB_shootdowns 4071 ± 6% -21.5% 3197 ± 2% interrupts.CPU78.TLB:TLB_shootdowns 4045 ± 5% -17.7% 3330 ± 7% interrupts.CPU79.TLB:TLB_shootdowns 454.00 ± 36% +51.5% 688.00 ± 13% interrupts.CPU80.RES:Rescheduling_interrupts 3963 ± 5% -18.5% 3229 ± 5% interrupts.CPU81.TLB:TLB_shootdowns 4202 ± 5% -25.3% 3139 ± 5% interrupts.CPU82.TLB:TLB_shootdowns 3843 ± 3% -15.6% 3242 interrupts.CPU83.TLB:TLB_shootdowns 4169 ± 5% -40.3% 2489 ± 57% interrupts.CPU86.TLB:TLB_shootdowns 4062 ± 6% -17.6% 3345 ± 2% interrupts.CPU87.TLB:TLB_shootdowns *************************************************************************************************** vm-snb-4G: qemu-system-x86_64 -enable-kvm -cpu SandyBridge -smp 2 -m 4G ========================================================================================= compiler/group/kconfig/rootfs/tbox_group/testcase: gcc-7/kselftests-02/x86_64-rhel-7.2/debian-x86_64-2018-04-03.cgz/vm-snb-4G/kernel_selftests commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | 4:12 -25% 1:12 kmsg.unregister_netdevice:waiting_for_veth_A-R1_to_become_free.Usage_count= 1:12 -8% :12 kmsg.veth0:Failed_to_cycle_device_veth0;route_tables_might_be_wrong :12 100% 12:12 kernel_selftests.memfd.run_fuse_test.sh.fail :12 8% 1:12 kernel_selftests.net.ip_defrag.sh.fail %stddev %change %stddev \ | \ 1.00 -100.0% 0.00 kernel_selftests.memfd.run_fuse_test.sh.pass 361623 +223.2% 1168751 meminfo.Active 361623 +223.2% 1168751 meminfo.Active(anon) 127769 ± 21% -94.8% 6662 ± 26% meminfo.CmaFree 1941240 -41.3% 1139756 meminfo.MemAvailable 1895394 -42.4% 1091194 ± 2% meminfo.MemFree 2137756 +37.6% 2941952 meminfo.Memused 881.70 ± 5% +52.8% 1347 ± 4% meminfo.max_used_kB *************************************************************************************************** lkp-bdw-ep3: 88 threads Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz with 64G memory ========================================================================================= class/compiler/cpufreq_governor/disk/kconfig/nr_threads/rootfs/tbox_group/testcase/testtime/ucode: pipe/gcc-7/performance/1HDD/x86_64-rhel-7.2/100%/debian-x86_64-2018-04-03.cgz/lkp-bdw-ep3/stress-ng/60s/0xb00002e commit: 9627026352 ("mm: page_cache_add_speculative(): refactoring") cdaa813278 ("mm/gup: track gup-pinned pages") 96270263521248d5 cdaa813278ddc616ee201eacda7 ---------------- --------------------------- fail:runs %reproduction fail:runs | | | :4 100% 4:4 dmesg.RIP:get_gup_pin_page :4 100% 4:4 dmesg.WARNING:at_mm/gup.c:#get_gup_pin_page %stddev %change %stddev \ | \ 360.22 -16.3% 301.59 stress-ng.time.elapsed_time 360.22 -16.3% 301.59 stress-ng.time.elapsed_time.max 8342 -1.3% 8234 stress-ng.time.percent_of_cpu_this_job_got 26890 -19.3% 21707 stress-ng.time.system_time 2.989e+08 -99.3% 2099213 stress-ng.vm-splice.ops 12010366 +10.0% 13211802 meminfo.Committed_AS 5812 ± 4% +18.0% 6856 ± 3% meminfo.max_used_kB 5.64 ± 2% +1.3 6.94 ± 2% mpstat.cpu.idle% 9.90 +1.8 11.68 ± 2% mpstat.cpu.usr% 107366 ± 93% -78.7% 22825 ± 44% numa-meminfo.node0.AnonPages 74249 ± 6% -13.8% 63969 ± 9% numa-meminfo.node0.SUnreclaim 5.90 ± 2% +22.8% 7.24 iostat.cpu.idle 84.23 -3.7% 81.11 iostat.cpu.system 9.87 +18.0% 11.64 ± 2% iostat.cpu.user 16463 +2.3% 16841 proc-vmstat.nr_kernel_stack 275921 ± 93% -50.8% 135706 ±150% proc-vmstat.numa_pte_updates 1000818 -14.3% 857522 proc-vmstat.pgfault 83.75 -3.6% 80.75 vmstat.cpu.sy 9.25 ± 4% +21.6% 11.25 ± 3% vmstat.cpu.us 1766858 +20.5% 2129192 vmstat.system.cs 185535 +1.5% 188310 vmstat.system.in 26840 ± 93% -78.8% 5697 ± 44% numa-vmstat.node0.nr_anon_pages 18563 ± 6% -13.8% 15992 ± 9% numa-vmstat.node0.nr_slab_unreclaimable 1.066e+09 ± 2% -28.2% 7.648e+08 ± 7% numa-vmstat.node0.numa_hit 1.066e+09 ± 2% -28.2% 7.648e+08 ± 7% numa-vmstat.node0.numa_local 9.301e+08 ± 2% -17.2% 7.701e+08 ± 6% numa-vmstat.node1.numa_hit 9.3e+08 ± 2% -17.2% 7.699e+08 ± 6% numa-vmstat.node1.numa_local 2.15 ± 2% -0.1 2.03 ± 4% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.do_idle.cpu_startup_entry.start_secondary 0.87 -0.1 0.81 perf-profile.calltrace.cycles-pp.avc_has_perm.file_has_perm.security_file_permission.vfs_read.ksys_read 1.35 -0.0 1.30 perf-profile.calltrace.cycles-pp.file_has_perm.security_file_permission.vfs_read.ksys_read.do_syscall_64 0.83 -0.0 0.80 perf-profile.calltrace.cycles-pp.avc_has_perm.file_has_perm.security_file_permission.vfs_write.ksys_write 1.31 -0.0 1.27 perf-profile.calltrace.cycles-pp.file_has_perm.security_file_permission.vfs_write.ksys_write.do_syscall_64 0.55 +0.0 0.57 perf-profile.calltrace.cycles-pp.__inode_security_revalidate.selinux_file_permission.security_file_permission.vfs_write.ksys_write 2.09 +0.1 2.15 perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.pipe_write.__vfs_write.vfs_write.ksys_write 2.62 ± 2% +0.1 2.68 perf-profile.calltrace.cycles-pp.pipe_wait.pipe_write.__vfs_write.vfs_write.ksys_write 2650 -1.3% 2616 turbostat.Avg_MHz 3.85 ± 2% +0.7 4.55 turbostat.C1% 0.54 ± 7% +0.2 0.70 ± 12% turbostat.C1E% 4.71 ± 2% +21.7% 5.73 turbostat.CPU%c1 77.50 ± 2% -14.2% 66.50 ± 2% turbostat.CoreTmp 67302470 -14.9% 57247592 turbostat.IRQ 81.75 ± 2% -13.1% 71.00 turbostat.PkgTmp 261.28 +3.9% 271.39 turbostat.PkgWatt 1285315 ± 8% +15.7% 1487427 ± 6% sched_debug.cfs_rq:/.MIN_vruntime.stddev 173175 -16.1% 145379 sched_debug.cfs_rq:/.exec_clock.avg 174188 -16.1% 146168 sched_debug.cfs_rq:/.exec_clock.max 171345 -16.0% 143970 sched_debug.cfs_rq:/.exec_clock.min 5689 ± 7% -27.5% 4124 ± 10% sched_debug.cfs_rq:/.load.min 168.54 ± 6% +12.3% 189.25 ± 6% sched_debug.cfs_rq:/.load_avg.max 1285315 ± 8% +15.7% 1487427 ± 6% sched_debug.cfs_rq:/.max_vruntime.stddev 16247293 -15.8% 13678528 sched_debug.cfs_rq:/.min_vruntime.avg 17621646 ± 2% -13.2% 15289779 ± 3% sched_debug.cfs_rq:/.min_vruntime.max 14916399 -18.2% 12204755 ± 3% sched_debug.cfs_rq:/.min_vruntime.min 0.13 ± 6% +21.7% 0.16 ± 8% sched_debug.cfs_rq:/.nr_running.stddev 4.96 ± 4% -20.3% 3.96 ± 9% sched_debug.cfs_rq:/.runnable_load_avg.min 5426 ± 8% -19.0% 4397 ± 11% sched_debug.cfs_rq:/.runnable_weight.min 10113 ± 21% -48.9% 5166 ± 13% sched_debug.cpu.avg_idle.min 222767 -13.4% 192825 sched_debug.cpu.clock.avg 222785 -13.4% 192841 sched_debug.cpu.clock.max 222745 -13.4% 192807 sched_debug.cpu.clock.min 11.90 ± 12% -17.7% 9.80 ± 12% sched_debug.cpu.clock.stddev 222767 -13.4% 192825 sched_debug.cpu.clock_task.avg 222785 -13.4% 192841 sched_debug.cpu.clock_task.max 222745 -13.4% 192807 sched_debug.cpu.clock_task.min 11.90 ± 12% -17.7% 9.80 ± 12% sched_debug.cpu.clock_task.stddev 5.21 ± 3% -12.9% 4.54 ± 3% sched_debug.cpu.cpu_load[0].min 4955 -19.3% 4001 sched_debug.cpu.curr->pid.avg 6909 -11.6% 6108 sched_debug.cpu.curr->pid.max 11634 ± 7% -10.9% 10363 sched_debug.cpu.load.avg 193768 -15.5% 163653 sched_debug.cpu.nr_load_updates.avg 197917 -15.6% 167117 sched_debug.cpu.nr_load_updates.max 191959 -15.6% 162097 sched_debug.cpu.nr_load_updates.min 0.50 ± 6% +10.8% 0.56 ± 6% sched_debug.cpu.nr_running.stddev 4644436 ± 2% -7.8% 4280609 ± 2% sched_debug.cpu.nr_switches.avg 4644663 ± 2% -7.8% 4280771 ± 2% sched_debug.cpu.sched_count.avg 409748 ± 4% -22.5% 317516 ± 3% sched_debug.cpu.sched_goidle.avg 572470 ± 11% -22.4% 444043 ± 13% sched_debug.cpu.sched_goidle.max 286861 ± 7% -22.8% 221401 ± 8% sched_debug.cpu.sched_goidle.min 222744 -13.4% 192807 sched_debug.cpu_clk 219014 -13.7% 189076 sched_debug.ktime 223427 -13.4% 193488 sched_debug.sched_clk 14.20 +15.8% 16.44 perf-stat.i.MPKI 3.166e+10 +17.1% 3.708e+10 perf-stat.i.branch-instructions 1.57 +0.0 1.59 perf-stat.i.branch-miss-rate% 4.847e+08 +16.6% 5.653e+08 ± 2% perf-stat.i.branch-misses 12.07 ± 23% -7.1 4.97 ± 55% perf-stat.i.cache-miss-rate% 9.997e+08 +18.7% 1.187e+09 perf-stat.i.cache-references 1782714 +20.5% 2148963 perf-stat.i.context-switches 4.01 -60.1% 1.60 perf-stat.i.cpi 269394 ± 5% +18.6% 319628 ± 2% perf-stat.i.cpu-migrations 0.19 ± 3% +0.0 0.22 perf-stat.i.dTLB-load-miss-rate% 47856194 ± 3% +19.6% 57235440 perf-stat.i.dTLB-load-misses 3.932e+10 +16.8% 4.592e+10 perf-stat.i.dTLB-loads 2.459e+10 +16.7% 2.871e+10 perf-stat.i.dTLB-stores 72.03 +1.3 73.35 perf-stat.i.iTLB-load-miss-rate% 92906511 ± 2% +16.5% 1.083e+08 ± 3% perf-stat.i.iTLB-load-misses 35482612 +12.6% 39969391 ± 2% perf-stat.i.iTLB-loads 1.521e+11 +17.0% 1.779e+11 perf-stat.i.instructions 28571 ± 3% +14.4% 32688 ± 6% perf-stat.i.instructions-per-iTLB-miss 0.64 +16.9% 0.75 perf-stat.i.ipc 2696 +1.9% 2746 perf-stat.i.minor-faults 115872 ± 4% +33.3% 154453 ± 11% perf-stat.i.node-loads 69.92 ± 5% -6.0 63.95 ± 2% perf-stat.i.node-store-miss-rate% 2696 +1.9% 2746 perf-stat.i.page-faults 1.53 -15.3% 1.30 perf-stat.overall.cpi 0.65 +18.1% 0.77 perf-stat.overall.ipc 3.153e+10 +16.9% 3.684e+10 perf-stat.ps.branch-instructions 4.827e+08 +16.3% 5.616e+08 ± 2% perf-stat.ps.branch-misses 9.972e+08 +18.4% 1.181e+09 perf-stat.ps.cache-references 1778082 +20.3% 2138597 perf-stat.ps.context-switches 2.319e+11 -1.2% 2.292e+11 perf-stat.ps.cpu-cycles 268859 ± 5% +18.4% 318278 ± 2% perf-stat.ps.cpu-migrations 47762824 ± 3% +19.4% 57005870 perf-stat.ps.dTLB-load-misses 3.916e+10 +16.5% 4.563e+10 perf-stat.ps.dTLB-loads 2.449e+10 +16.5% 2.852e+10 perf-stat.ps.dTLB-stores 92569828 ± 2% +16.2% 1.076e+08 ± 3% perf-stat.ps.iTLB-load-misses 35350893 +12.4% 39724157 ± 2% perf-stat.ps.iTLB-loads 1.515e+11 +16.7% 1.767e+11 perf-stat.ps.instructions 2690 +1.6% 2733 perf-stat.ps.minor-faults 115582 ± 4% +33.0% 153669 ± 11% perf-stat.ps.node-loads 2690 +1.6% 2733 perf-stat.ps.page-faults 5.463e+13 -2.2% 5.344e+13 perf-stat.total.instructions 423.00 -16.0% 355.50 interrupts.9:IO-APIC.9-fasteoi.acpi 334285 -13.8% 288189 interrupts.CAL:Function_call_interrupts 3718 ± 5% -11.2% 3301 interrupts.CPU0.CAL:Function_call_interrupts 721763 -16.2% 605183 interrupts.CPU0.LOC:Local_timer_interrupts 25452 ± 12% +29.7% 33009 ± 11% interrupts.CPU0.RES:Rescheduling_interrupts 423.00 -16.0% 355.50 interrupts.CPU1.9:IO-APIC.9-fasteoi.acpi 3841 -13.6% 3317 interrupts.CPU1.CAL:Function_call_interrupts 721450 -16.2% 604748 interrupts.CPU1.LOC:Local_timer_interrupts 3776 ± 2% -12.2% 3317 interrupts.CPU10.CAL:Function_call_interrupts 722124 -16.2% 604931 interrupts.CPU10.LOC:Local_timer_interrupts 22581 ± 13% +44.5% 32630 ± 25% interrupts.CPU10.RES:Rescheduling_interrupts 3835 -13.4% 3321 interrupts.CPU11.CAL:Function_call_interrupts 721997 -16.2% 605063 interrupts.CPU11.LOC:Local_timer_interrupts 21969 ± 12% +39.6% 30662 ± 10% interrupts.CPU11.RES:Rescheduling_interrupts 3828 -12.7% 3340 interrupts.CPU12.CAL:Function_call_interrupts 721760 -16.2% 604828 interrupts.CPU12.LOC:Local_timer_interrupts 3790 -12.8% 3303 interrupts.CPU13.CAL:Function_call_interrupts 721659 -16.2% 604978 interrupts.CPU13.LOC:Local_timer_interrupts 3796 -12.6% 3317 interrupts.CPU14.CAL:Function_call_interrupts 721830 -16.2% 605051 interrupts.CPU14.LOC:Local_timer_interrupts 3745 ± 2% -11.6% 3313 interrupts.CPU15.CAL:Function_call_interrupts 721696 -16.1% 605247 interrupts.CPU15.LOC:Local_timer_interrupts 3796 ± 2% -13.1% 3298 interrupts.CPU16.CAL:Function_call_interrupts 721615 -16.1% 605304 interrupts.CPU16.LOC:Local_timer_interrupts 721686 -16.2% 604977 interrupts.CPU17.LOC:Local_timer_interrupts 23557 ± 11% +31.8% 31037 ± 17% interrupts.CPU17.RES:Rescheduling_interrupts 3841 -13.9% 3308 interrupts.CPU18.CAL:Function_call_interrupts 721620 -16.2% 604958 interrupts.CPU18.LOC:Local_timer_interrupts 22764 ± 17% +33.8% 30451 ± 8% interrupts.CPU18.RES:Rescheduling_interrupts 3790 ± 2% -12.8% 3304 interrupts.CPU19.CAL:Function_call_interrupts 721479 -16.1% 604999 interrupts.CPU19.LOC:Local_timer_interrupts 3845 -14.0% 3307 interrupts.CPU2.CAL:Function_call_interrupts 721284 -16.1% 605094 interrupts.CPU2.LOC:Local_timer_interrupts 22880 ± 19% +42.0% 32484 ± 12% interrupts.CPU2.RES:Rescheduling_interrupts 721360 -16.1% 604946 interrupts.CPU20.LOC:Local_timer_interrupts 21116 ± 18% +39.1% 29366 ± 17% interrupts.CPU20.RES:Rescheduling_interrupts 3811 -13.5% 3297 interrupts.CPU21.CAL:Function_call_interrupts 721473 -16.2% 604796 interrupts.CPU21.LOC:Local_timer_interrupts 3827 -15.3% 3241 ± 3% interrupts.CPU22.CAL:Function_call_interrupts 721210 -16.2% 604237 interrupts.CPU22.LOC:Local_timer_interrupts 721987 -16.2% 604725 interrupts.CPU23.LOC:Local_timer_interrupts 3821 -19.4% 3081 ± 6% interrupts.CPU24.CAL:Function_call_interrupts 721544 -16.2% 604685 interrupts.CPU24.LOC:Local_timer_interrupts 3786 -15.6% 3194 ± 3% interrupts.CPU25.CAL:Function_call_interrupts 721135 -16.1% 604853 interrupts.CPU25.LOC:Local_timer_interrupts 3827 -15.2% 3245 ± 2% interrupts.CPU26.CAL:Function_call_interrupts 721813 -16.2% 605191 interrupts.CPU26.LOC:Local_timer_interrupts 3821 -25.4% 2851 ± 26% interrupts.CPU27.CAL:Function_call_interrupts 721056 -16.1% 605078 interrupts.CPU27.LOC:Local_timer_interrupts 3818 -15.1% 3241 ± 3% interrupts.CPU28.CAL:Function_call_interrupts 721295 -16.1% 604867 interrupts.CPU28.LOC:Local_timer_interrupts 3839 -14.2% 3293 interrupts.CPU29.CAL:Function_call_interrupts 721791 -16.2% 604635 interrupts.CPU29.LOC:Local_timer_interrupts 3818 -13.1% 3318 interrupts.CPU3.CAL:Function_call_interrupts 721555 -16.2% 604385 interrupts.CPU3.LOC:Local_timer_interrupts 22807 ± 12% +32.9% 30316 ± 13% interrupts.CPU3.RES:Rescheduling_interrupts 3829 -14.3% 3280 interrupts.CPU30.CAL:Function_call_interrupts 721877 -16.2% 604644 interrupts.CPU30.LOC:Local_timer_interrupts 29149 ± 5% -15.8% 24552 ± 10% interrupts.CPU30.RES:Rescheduling_interrupts 3838 -13.9% 3304 interrupts.CPU31.CAL:Function_call_interrupts 721927 -16.2% 604849 interrupts.CPU31.LOC:Local_timer_interrupts 3836 -14.2% 3290 interrupts.CPU32.CAL:Function_call_interrupts 720992 -16.1% 604786 interrupts.CPU32.LOC:Local_timer_interrupts 3826 -14.0% 3292 interrupts.CPU33.CAL:Function_call_interrupts 721853 -16.2% 604700 interrupts.CPU33.LOC:Local_timer_interrupts 3809 -13.8% 3283 interrupts.CPU34.CAL:Function_call_interrupts 721742 -16.2% 605038 interrupts.CPU34.LOC:Local_timer_interrupts 3833 -14.0% 3296 interrupts.CPU35.CAL:Function_call_interrupts 721732 -16.2% 604841 interrupts.CPU35.LOC:Local_timer_interrupts 3700 ± 6% -12.7% 3229 ± 3% interrupts.CPU36.CAL:Function_call_interrupts 721598 -16.2% 604849 interrupts.CPU36.LOC:Local_timer_interrupts 3835 -14.1% 3293 interrupts.CPU37.CAL:Function_call_interrupts 721245 -16.2% 604498 interrupts.CPU37.LOC:Local_timer_interrupts 3831 -13.8% 3301 interrupts.CPU38.CAL:Function_call_interrupts 721366 -16.1% 605049 interrupts.CPU38.LOC:Local_timer_interrupts 3843 -15.3% 3253 ± 2% interrupts.CPU39.CAL:Function_call_interrupts 721951 -16.2% 604693 interrupts.CPU39.LOC:Local_timer_interrupts 3823 -13.2% 3316 interrupts.CPU4.CAL:Function_call_interrupts 721523 -16.2% 604953 interrupts.CPU4.LOC:Local_timer_interrupts 22707 ± 16% +33.2% 30241 ± 3% interrupts.CPU4.RES:Rescheduling_interrupts 3840 -14.2% 3294 interrupts.CPU40.CAL:Function_call_interrupts 721636 -16.2% 604810 interrupts.CPU40.LOC:Local_timer_interrupts 3847 -18.8% 3122 ± 9% interrupts.CPU41.CAL:Function_call_interrupts 721928 -16.2% 604884 interrupts.CPU41.LOC:Local_timer_interrupts 3839 -14.3% 3290 interrupts.CPU42.CAL:Function_call_interrupts 721753 -16.2% 604811 interrupts.CPU42.LOC:Local_timer_interrupts 3778 -13.8% 3255 interrupts.CPU43.CAL:Function_call_interrupts 721315 -16.2% 604691 interrupts.CPU43.LOC:Local_timer_interrupts 3841 -13.7% 3314 interrupts.CPU44.CAL:Function_call_interrupts 721306 -16.2% 604640 interrupts.CPU44.LOC:Local_timer_interrupts 22357 ± 16% +47.9% 33078 ± 15% interrupts.CPU44.RES:Rescheduling_interrupts 3832 -15.0% 3256 ± 3% interrupts.CPU45.CAL:Function_call_interrupts 721470 -16.2% 604760 interrupts.CPU45.LOC:Local_timer_interrupts 3846 -18.1% 3149 ± 9% interrupts.CPU46.CAL:Function_call_interrupts 721294 -16.2% 604606 interrupts.CPU46.LOC:Local_timer_interrupts 22430 ± 19% +39.5% 31283 ± 10% interrupts.CPU46.RES:Rescheduling_interrupts 3795 -12.5% 3320 interrupts.CPU47.CAL:Function_call_interrupts 721274 -16.2% 604302 interrupts.CPU47.LOC:Local_timer_interrupts 22255 ± 13% +29.9% 28907 ± 5% interrupts.CPU47.RES:Rescheduling_interrupts 3841 -14.2% 3295 interrupts.CPU48.CAL:Function_call_interrupts 721357 -16.2% 604782 interrupts.CPU48.LOC:Local_timer_interrupts 20959 ± 17% +36.5% 28600 ± 7% interrupts.CPU48.RES:Rescheduling_interrupts 3835 -13.3% 3324 interrupts.CPU49.CAL:Function_call_interrupts 721311 -16.1% 605127 interrupts.CPU49.LOC:Local_timer_interrupts 22089 ± 15% +33.1% 29406 ± 11% interrupts.CPU49.RES:Rescheduling_interrupts 3842 -14.0% 3305 interrupts.CPU5.CAL:Function_call_interrupts 721561 -16.1% 605092 interrupts.CPU5.LOC:Local_timer_interrupts 24272 ± 14% +25.3% 30413 ± 11% interrupts.CPU5.RES:Rescheduling_interrupts 3848 -13.7% 3319 interrupts.CPU50.CAL:Function_call_interrupts 721251 -16.1% 604921 interrupts.CPU50.LOC:Local_timer_interrupts 3827 -13.3% 3318 interrupts.CPU51.CAL:Function_call_interrupts 721474 -16.2% 604842 interrupts.CPU51.LOC:Local_timer_interrupts 3812 -12.5% 3337 interrupts.CPU52.CAL:Function_call_interrupts 721284 -16.1% 604942 interrupts.CPU52.LOC:Local_timer_interrupts 22236 ± 10% +33.3% 29642 ± 8% interrupts.CPU52.RES:Rescheduling_interrupts 721452 -16.1% 605377 interrupts.CPU53.LOC:Local_timer_interrupts 21926 ± 15% +38.2% 30305 ± 12% interrupts.CPU53.RES:Rescheduling_interrupts 721468 -16.1% 605249 interrupts.CPU54.LOC:Local_timer_interrupts 22522 ± 13% +36.4% 30717 ± 17% interrupts.CPU54.RES:Rescheduling_interrupts 3842 -13.9% 3307 interrupts.CPU55.CAL:Function_call_interrupts 721620 -16.2% 604986 interrupts.CPU55.LOC:Local_timer_interrupts 22201 ± 16% +31.7% 29239 ± 11% interrupts.CPU55.RES:Rescheduling_interrupts 3811 -13.3% 3303 interrupts.CPU56.CAL:Function_call_interrupts 721916 -16.2% 605011 interrupts.CPU56.LOC:Local_timer_interrupts 721636 -16.2% 604961 interrupts.CPU57.LOC:Local_timer_interrupts 3841 -13.9% 3309 interrupts.CPU58.CAL:Function_call_interrupts 721637 -16.1% 605154 interrupts.CPU58.LOC:Local_timer_interrupts 3795 -12.7% 3314 interrupts.CPU59.CAL:Function_call_interrupts 721926 -16.1% 605366 interrupts.CPU59.LOC:Local_timer_interrupts 3841 -17.0% 3189 ± 7% interrupts.CPU6.CAL:Function_call_interrupts 721537 -16.1% 605025 interrupts.CPU6.LOC:Local_timer_interrupts 21627 ± 16% +36.9% 29599 ± 15% interrupts.CPU6.RES:Rescheduling_interrupts 3839 -13.5% 3319 interrupts.CPU60.CAL:Function_call_interrupts 721689 -16.2% 604938 interrupts.CPU60.LOC:Local_timer_interrupts 22724 ± 12% +31.4% 29852 ± 13% interrupts.CPU60.RES:Rescheduling_interrupts 721538 -16.1% 605018 interrupts.CPU61.LOC:Local_timer_interrupts 22521 ± 15% +33.2% 29995 ± 19% interrupts.CPU61.RES:Rescheduling_interrupts 3795 ± 2% -12.7% 3313 interrupts.CPU62.CAL:Function_call_interrupts 721565 -16.2% 604946 interrupts.CPU62.LOC:Local_timer_interrupts 22054 ± 17% +38.7% 30598 ± 11% interrupts.CPU62.RES:Rescheduling_interrupts 721884 -16.2% 605030 interrupts.CPU63.LOC:Local_timer_interrupts 3772 ± 3% -12.3% 3309 interrupts.CPU64.CAL:Function_call_interrupts 721981 -16.2% 605040 interrupts.CPU64.LOC:Local_timer_interrupts 21437 ± 20% +33.6% 28639 ± 19% interrupts.CPU64.RES:Rescheduling_interrupts 3779 -19.3% 3051 ± 14% interrupts.CPU65.CAL:Function_call_interrupts 721390 -16.1% 605132 interrupts.CPU65.LOC:Local_timer_interrupts 3825 -13.5% 3309 interrupts.CPU66.CAL:Function_call_interrupts 722064 -16.3% 604606 interrupts.CPU66.LOC:Local_timer_interrupts 3819 -26.5% 2808 ± 24% interrupts.CPU67.CAL:Function_call_interrupts 721439 -16.2% 604744 interrupts.CPU67.LOC:Local_timer_interrupts 3821 -13.5% 3306 interrupts.CPU68.CAL:Function_call_interrupts 722060 -16.3% 604673 interrupts.CPU68.LOC:Local_timer_interrupts 3821 -13.7% 3297 interrupts.CPU69.CAL:Function_call_interrupts 722106 -16.3% 604542 interrupts.CPU69.LOC:Local_timer_interrupts 3842 -14.2% 3297 interrupts.CPU7.CAL:Function_call_interrupts 721363 -16.1% 605295 interrupts.CPU7.LOC:Local_timer_interrupts 22221 ± 16% +40.9% 31317 ± 19% interrupts.CPU7.RES:Rescheduling_interrupts 3826 -13.6% 3305 interrupts.CPU70.CAL:Function_call_interrupts 721764 -16.2% 604627 interrupts.CPU70.LOC:Local_timer_interrupts 3809 -13.5% 3295 interrupts.CPU71.CAL:Function_call_interrupts 722080 -16.2% 604781 interrupts.CPU71.LOC:Local_timer_interrupts 3818 -14.2% 3274 ± 2% interrupts.CPU72.CAL:Function_call_interrupts 721916 -16.2% 604918 interrupts.CPU72.LOC:Local_timer_interrupts 27835 ± 14% -17.6% 22945 ± 10% interrupts.CPU72.RES:Rescheduling_interrupts 3813 -13.2% 3308 interrupts.CPU73.CAL:Function_call_interrupts 721747 -16.2% 604707 interrupts.CPU73.LOC:Local_timer_interrupts 721746 -16.2% 604838 interrupts.CPU74.LOC:Local_timer_interrupts 3832 -14.7% 3269 ± 2% interrupts.CPU75.CAL:Function_call_interrupts 721781 -16.2% 604619 interrupts.CPU75.LOC:Local_timer_interrupts 3779 -12.4% 3310 interrupts.CPU76.CAL:Function_call_interrupts 720970 -16.1% 604618 interrupts.CPU76.LOC:Local_timer_interrupts 3809 ± 2% -15.7% 3211 ± 4% interrupts.CPU77.CAL:Function_call_interrupts 722062 -16.3% 604188 interrupts.CPU77.LOC:Local_timer_interrupts 3843 -14.6% 3284 ± 2% interrupts.CPU78.CAL:Function_call_interrupts 721715 -16.2% 604636 interrupts.CPU78.LOC:Local_timer_interrupts 3840 -13.9% 3306 interrupts.CPU79.CAL:Function_call_interrupts 721613 -16.2% 604526 interrupts.CPU79.LOC:Local_timer_interrupts 3763 ± 4% -11.5% 3331 interrupts.CPU8.CAL:Function_call_interrupts 721367 -16.1% 605107 interrupts.CPU8.LOC:Local_timer_interrupts 22471 ± 17% +38.2% 31048 ± 9% interrupts.CPU8.RES:Rescheduling_interrupts 3830 -13.9% 3298 interrupts.CPU80.CAL:Function_call_interrupts 721678 -16.2% 604714 interrupts.CPU80.LOC:Local_timer_interrupts 3776 ± 3% -12.5% 3306 interrupts.CPU81.CAL:Function_call_interrupts 721263 -16.1% 604803 interrupts.CPU81.LOC:Local_timer_interrupts 3837 -13.9% 3305 interrupts.CPU82.CAL:Function_call_interrupts 721961 -16.2% 604781 interrupts.CPU82.LOC:Local_timer_interrupts 3849 -14.8% 3280 interrupts.CPU83.CAL:Function_call_interrupts 721888 -16.2% 604619 interrupts.CPU83.LOC:Local_timer_interrupts 3802 -13.2% 3300 interrupts.CPU84.CAL:Function_call_interrupts 721937 -16.2% 604653 interrupts.CPU84.LOC:Local_timer_interrupts 3843 -13.8% 3313 interrupts.CPU85.CAL:Function_call_interrupts 721834 -16.2% 604711 interrupts.CPU85.LOC:Local_timer_interrupts 3839 -14.3% 3291 interrupts.CPU86.CAL:Function_call_interrupts 721871 -16.2% 604809 interrupts.CPU86.LOC:Local_timer_interrupts 3847 -14.6% 3287 interrupts.CPU87.CAL:Function_call_interrupts 722149 -16.3% 604789 interrupts.CPU87.LOC:Local_timer_interrupts 3838 -14.1% 3298 interrupts.CPU9.CAL:Function_call_interrupts 721253 -16.2% 604595 interrupts.CPU9.LOC:Local_timer_interrupts 22535 ± 19% +36.1% 30680 ± 11% interrupts.CPU9.RES:Rescheduling_interrupts 63503101 -16.2% 53227032 interrupts.LOC:Local_timer_interrupts 310.00 ± 3% +11.0% 344.00 ± 7% interrupts.TLB:TLB_shootdowns 20431 ± 15% -19.7% 16415 ± 4% softirqs.CPU0.RCU 34582 ± 2% +17.5% 40639 ± 8% softirqs.CPU0.SCHED 133704 -12.9% 116517 ± 3% softirqs.CPU0.TIMER 20017 ± 16% -18.6% 16300 softirqs.CPU1.RCU 32010 ± 3% +17.3% 37561 ± 8% softirqs.CPU1.SCHED 132385 ± 2% -14.8% 112812 softirqs.CPU1.TIMER 19743 ± 16% -20.2% 15764 ± 3% softirqs.CPU10.RCU 30881 ± 3% +20.6% 37253 ± 10% softirqs.CPU10.SCHED 130801 ± 2% -8.5% 119711 ± 6% softirqs.CPU10.TIMER 30607 ± 2% +20.4% 36850 ± 8% softirqs.CPU11.SCHED 133304 ± 2% -11.9% 117463 ± 2% softirqs.CPU11.TIMER 30848 +20.1% 37052 ± 9% softirqs.CPU12.SCHED 131208 -11.5% 116174 ± 3% softirqs.CPU12.TIMER 22042 ± 17% -28.7% 15719 ± 5% softirqs.CPU13.RCU 30520 ± 5% +21.7% 37133 ± 11% softirqs.CPU13.SCHED 135598 ± 4% -15.4% 114708 ± 2% softirqs.CPU13.TIMER 31132 ± 4% +15.4% 35911 ± 10% softirqs.CPU14.SCHED 148384 ± 17% -22.8% 114512 softirqs.CPU14.TIMER 30303 +20.6% 36560 ± 10% softirqs.CPU15.SCHED 133852 -14.1% 115008 softirqs.CPU15.TIMER 31199 ± 2% +18.5% 36961 ± 8% softirqs.CPU16.SCHED 133767 ± 2% -13.1% 116191 softirqs.CPU16.TIMER 19872 ± 17% -17.2% 16453 ± 2% softirqs.CPU17.RCU 30661 ± 2% +19.2% 36554 ± 9% softirqs.CPU17.SCHED 133274 ± 2% -15.1% 113122 ± 2% softirqs.CPU17.TIMER 20088 ± 15% -17.5% 16582 ± 2% softirqs.CPU18.RCU 31180 ± 3% +15.6% 36055 ± 9% softirqs.CPU18.SCHED 134796 ± 4% -15.7% 113667 softirqs.CPU18.TIMER 22481 ± 20% -26.8% 16451 ± 2% softirqs.CPU19.RCU 30809 ± 2% +19.5% 36811 ± 7% softirqs.CPU19.SCHED 132271 ± 2% -12.4% 115843 ± 3% softirqs.CPU19.TIMER 19797 ± 16% -17.5% 16326 ± 2% softirqs.CPU2.RCU 30922 ± 3% +21.6% 37612 ± 8% softirqs.CPU2.SCHED 25248 ± 29% -24.3% 19119 ± 24% softirqs.CPU20.RCU 29985 ± 2% +22.1% 36608 ± 8% softirqs.CPU20.SCHED 128772 ± 2% -10.9% 114692 softirqs.CPU20.TIMER 19935 ± 17% -19.2% 16110 ± 3% softirqs.CPU21.RCU 31495 ± 3% +16.4% 36651 ± 7% softirqs.CPU21.SCHED 130837 ± 3% -13.2% 113573 ± 2% softirqs.CPU21.TIMER 40257 ± 5% -13.7% 34739 ± 7% softirqs.CPU22.SCHED 133177 ± 2% -16.6% 111009 ± 3% softirqs.CPU22.TIMER 40524 ± 5% -17.6% 33383 ± 9% softirqs.CPU23.SCHED 133547 ± 4% -18.1% 109369 softirqs.CPU23.TIMER 39849 ± 3% -15.3% 33755 ± 9% softirqs.CPU24.SCHED 131526 ± 2% -17.7% 108249 ± 2% softirqs.CPU24.TIMER 39996 ± 4% -13.6% 34576 ± 6% softirqs.CPU25.SCHED 131627 ± 3% -16.1% 110442 ± 3% softirqs.CPU25.TIMER 39898 ± 5% -14.9% 33947 ± 7% softirqs.CPU26.SCHED 132681 -16.5% 110841 ± 4% softirqs.CPU26.TIMER 40079 ± 7% -15.6% 33826 ± 6% softirqs.CPU27.SCHED 133577 -18.1% 109462 ± 2% softirqs.CPU27.TIMER 40228 ± 5% -15.4% 34036 ± 6% softirqs.CPU28.SCHED 131075 ± 3% -17.9% 107666 ± 2% softirqs.CPU28.TIMER 39312 ± 3% -14.8% 33499 ± 8% softirqs.CPU29.SCHED 134423 ± 3% -18.4% 109702 ± 3% softirqs.CPU29.TIMER 19384 ± 17% -17.8% 15929 softirqs.CPU3.RCU 30436 ± 3% +21.2% 36897 ± 11% softirqs.CPU3.SCHED 132319 ± 3% -13.4% 114525 ± 2% softirqs.CPU3.TIMER 22908 ± 16% -22.4% 17781 ± 3% softirqs.CPU30.RCU 39762 ± 3% -15.0% 33803 ± 8% softirqs.CPU30.SCHED 134269 ± 2% -17.7% 110531 ± 2% softirqs.CPU30.TIMER 41024 ± 5% -17.4% 33894 ± 7% softirqs.CPU31.SCHED 149024 ± 17% -24.6% 112375 ± 3% softirqs.CPU31.TIMER 39978 ± 4% -15.5% 33774 ± 7% softirqs.CPU32.SCHED 131148 -17.2% 108564 ± 3% softirqs.CPU32.TIMER 39979 ± 4% -13.8% 34476 ± 9% softirqs.CPU33.SCHED 134094 ± 4% -18.1% 109836 ± 2% softirqs.CPU33.TIMER 22786 ± 16% -22.9% 17569 ± 4% softirqs.CPU34.RCU 40552 ± 4% -17.3% 33554 ± 7% softirqs.CPU34.SCHED 131088 ± 2% -17.0% 108798 ± 3% softirqs.CPU34.TIMER 39532 ± 3% -14.1% 33963 ± 7% softirqs.CPU35.SCHED 132211 ± 2% -15.9% 111137 ± 3% softirqs.CPU35.TIMER 38671 ± 4% -13.2% 33557 ± 9% softirqs.CPU36.SCHED 147868 ± 17% -27.0% 107982 softirqs.CPU36.TIMER 23249 ± 15% -22.1% 18100 ± 5% softirqs.CPU37.RCU 40190 ± 5% -15.3% 34023 ± 7% softirqs.CPU37.SCHED 132884 ± 3% -15.9% 111703 ± 2% softirqs.CPU37.TIMER 22745 ± 16% -23.2% 17469 ± 4% softirqs.CPU38.RCU 40569 ± 5% -16.3% 33950 ± 9% softirqs.CPU38.SCHED 135133 -18.3% 110441 ± 2% softirqs.CPU38.TIMER 39935 ± 5% -15.3% 33825 ± 8% softirqs.CPU39.SCHED 133753 ± 2% -17.5% 110412 softirqs.CPU39.TIMER 30946 ± 4% +17.7% 36409 ± 9% softirqs.CPU4.SCHED 130762 ± 2% -13.6% 113024 ± 2% softirqs.CPU4.TIMER 40335 ± 3% -16.4% 33703 ± 9% softirqs.CPU40.SCHED 130983 ± 2% -15.6% 110597 ± 2% softirqs.CPU40.TIMER 40423 ± 5% -17.1% 33525 ± 8% softirqs.CPU41.SCHED 132224 ± 4% -18.0% 108363 softirqs.CPU41.TIMER 40723 ± 4% -17.3% 33682 ± 7% softirqs.CPU42.SCHED 135049 -17.8% 110952 ± 3% softirqs.CPU42.TIMER 39349 ± 3% -14.8% 33536 ± 8% softirqs.CPU43.SCHED 134203 ± 2% -17.6% 110610 ± 2% softirqs.CPU43.TIMER 17084 ± 18% -18.0% 14000 ± 3% softirqs.CPU44.RCU 30946 +22.1% 37787 ± 9% softirqs.CPU44.SCHED 130568 -13.3% 113197 ± 3% softirqs.CPU44.TIMER 20252 ± 15% -19.8% 16238 ± 2% softirqs.CPU45.RCU 31638 ± 4% +16.3% 36780 ± 9% softirqs.CPU45.SCHED 132331 ± 2% -14.9% 112550 softirqs.CPU45.TIMER 24973 ± 13% -16.0% 20983 ± 2% softirqs.CPU46.RCU 30862 +22.1% 37672 ± 6% softirqs.CPU46.SCHED 137074 ± 3% -11.2% 121691 ± 4% softirqs.CPU46.TIMER 19750 ± 17% -18.6% 16079 ± 2% softirqs.CPU47.RCU 30289 ± 3% +21.4% 36762 ± 9% softirqs.CPU47.SCHED 130548 ± 3% -12.7% 113914 softirqs.CPU47.TIMER 19693 ± 16% -17.0% 16336 ± 4% softirqs.CPU48.RCU 30593 ± 5% +19.9% 36684 ± 9% softirqs.CPU48.SCHED 130499 ± 3% -13.6% 112753 ± 2% softirqs.CPU48.TIMER 30877 ± 2% +19.9% 37018 ± 9% softirqs.CPU49.SCHED 130875 ± 4% -13.0% 113833 softirqs.CPU49.TIMER 22365 ± 36% -27.9% 16131 softirqs.CPU5.RCU 30822 +20.2% 37048 ± 9% softirqs.CPU5.SCHED 130397 ± 3% -11.8% 115024 softirqs.CPU5.TIMER 22536 ± 37% -26.9% 16472 ± 4% softirqs.CPU50.RCU 30543 +20.3% 36752 ± 9% softirqs.CPU50.SCHED 133918 ± 3% -13.5% 115860 ± 4% softirqs.CPU50.TIMER 131702 ± 2% -9.6% 118995 ± 6% softirqs.CPU51.TIMER 20176 ± 17% -21.1% 15915 ± 3% softirqs.CPU52.RCU 30631 ± 3% +22.2% 37419 ± 7% softirqs.CPU52.SCHED 131629 -13.5% 113828 softirqs.CPU52.TIMER 31078 +21.4% 37736 ± 9% softirqs.CPU53.SCHED 133106 ± 3% -14.2% 114257 ± 3% softirqs.CPU53.TIMER 30982 ± 4% +19.3% 36962 ± 9% softirqs.CPU54.SCHED 25187 ± 31% -24.8% 18932 ± 27% softirqs.CPU55.RCU 30622 ± 3% +19.6% 36613 ± 9% softirqs.CPU55.SCHED 19796 ± 14% -18.3% 16167 ± 3% softirqs.CPU56.RCU 30828 +18.8% 36619 ± 9% softirqs.CPU56.SCHED 130194 ± 2% -12.1% 114503 ± 2% softirqs.CPU56.TIMER 30577 ± 5% +21.1% 37016 ± 10% softirqs.CPU57.SCHED 147818 ± 17% -24.6% 111515 ± 2% softirqs.CPU57.TIMER 25343 ± 33% -36.6% 16078 ± 4% softirqs.CPU58.RCU 31181 ± 2% +16.6% 36349 ± 10% softirqs.CPU58.SCHED 139070 ± 7% -16.4% 116283 ± 2% softirqs.CPU58.TIMER 19741 ± 16% -15.3% 16722 ± 5% softirqs.CPU59.RCU 30636 ± 2% +20.1% 36788 ± 9% softirqs.CPU59.SCHED 133696 -14.0% 114933 softirqs.CPU59.TIMER 30639 ± 2% +19.3% 36549 ± 9% softirqs.CPU6.SCHED 31080 +19.0% 36975 ± 7% softirqs.CPU60.SCHED 133580 ± 3% -13.5% 115606 softirqs.CPU60.TIMER 19131 ± 16% -16.9% 15898 ± 2% softirqs.CPU61.RCU 30495 ± 3% +20.2% 36664 ± 10% softirqs.CPU61.SCHED 133103 -15.3% 112675 softirqs.CPU61.TIMER 31019 ± 2% +18.8% 36849 ± 8% softirqs.CPU62.SCHED 134432 ± 4% -15.8% 113144 softirqs.CPU62.TIMER 30882 ± 2% +16.9% 36105 ± 8% softirqs.CPU63.SCHED 131772 ± 2% -12.7% 115005 ± 3% softirqs.CPU63.TIMER 19316 ± 16% -18.8% 15694 ± 4% softirqs.CPU64.RCU 30286 ± 2% +20.5% 36497 ± 8% softirqs.CPU64.SCHED 128582 ± 2% -10.1% 115656 ± 4% softirqs.CPU64.TIMER 19223 ± 17% -18.1% 15749 ± 3% softirqs.CPU65.RCU 30985 ± 2% +18.6% 36751 ± 8% softirqs.CPU65.SCHED 130355 ± 3% -13.0% 113407 ± 2% softirqs.CPU65.TIMER 40027 ± 4% -14.7% 34153 ± 7% softirqs.CPU66.SCHED 131990 ± 3% -16.9% 109630 ± 3% softirqs.CPU66.TIMER 40103 ± 5% -16.4% 33528 ± 7% softirqs.CPU67.SCHED 133065 ± 4% -18.5% 108445 softirqs.CPU67.TIMER 39282 -14.5% 33576 ± 9% softirqs.CPU68.SCHED 131993 ± 2% -18.9% 107006 ± 2% softirqs.CPU68.TIMER 131342 ± 3% -16.3% 109974 ± 3% softirqs.CPU69.TIMER 19684 ± 18% -16.4% 16446 ± 4% softirqs.CPU7.RCU 30804 ± 2% +19.7% 36877 ± 11% softirqs.CPU7.SCHED 39499 ± 4% -14.1% 33939 ± 7% softirqs.CPU70.SCHED 132374 -17.3% 109508 ± 4% softirqs.CPU70.TIMER 39816 ± 4% -13.6% 34387 ± 6% softirqs.CPU71.SCHED 133540 -18.0% 109519 ± 2% softirqs.CPU71.TIMER 40073 ± 4% -15.8% 33726 ± 6% softirqs.CPU72.SCHED 132172 ± 4% -18.5% 107744 softirqs.CPU72.TIMER 39089 ± 3% -14.1% 33596 ± 7% softirqs.CPU73.SCHED 131584 -16.4% 110040 ± 3% softirqs.CPU73.TIMER 40315 ± 4% -16.5% 33655 ± 7% softirqs.CPU74.SCHED 133455 -17.6% 109941 softirqs.CPU74.TIMER 40427 ± 5% -16.7% 33659 ± 8% softirqs.CPU75.SCHED 137065 ± 5% -18.6% 111601 ± 3% softirqs.CPU75.TIMER 39577 ± 3% -14.0% 34031 ± 7% softirqs.CPU76.SCHED 130476 -16.6% 108755 ± 4% softirqs.CPU76.TIMER 40386 ± 5% -14.9% 34356 ± 7% softirqs.CPU77.SCHED 134414 ± 4% -18.8% 109199 ± 3% softirqs.CPU77.TIMER 40179 ± 5% -16.2% 33656 ± 7% softirqs.CPU78.SCHED 131087 ± 2% -17.4% 108250 ± 3% softirqs.CPU78.TIMER 39263 ± 3% -14.2% 33686 ± 7% softirqs.CPU79.SCHED 130805 ± 2% -16.3% 109502 ± 2% softirqs.CPU79.TIMER 19896 ± 16% -19.7% 15987 ± 2% softirqs.CPU8.RCU 30563 ± 3% +24.0% 37906 ± 7% softirqs.CPU8.SCHED 131375 ± 2% -12.8% 114531 softirqs.CPU8.TIMER 39276 ± 2% -14.9% 33416 ± 9% softirqs.CPU80.SCHED 138042 ± 7% -21.9% 107876 ± 2% softirqs.CPU80.TIMER 21626 ± 17% -24.1% 16414 ± 4% softirqs.CPU81.RCU 40018 ± 4% -15.6% 33781 ± 7% softirqs.CPU81.SCHED 134117 ± 4% -16.8% 111583 ± 3% softirqs.CPU81.TIMER 134747 -18.0% 110540 ± 3% softirqs.CPU82.TIMER 133707 ± 2% -17.6% 110191 softirqs.CPU83.TIMER 40079 ± 4% -16.3% 33556 ± 10% softirqs.CPU84.SCHED 130536 ± 2% -15.8% 109959 softirqs.CPU84.TIMER 40427 ± 5% -16.9% 33585 ± 7% softirqs.CPU85.SCHED 131805 ± 4% -18.4% 107549 softirqs.CPU85.TIMER 40785 ± 5% -17.5% 33639 ± 7% softirqs.CPU86.SCHED 134917 ± 2% -18.1% 110473 ± 3% softirqs.CPU86.TIMER 39496 ± 3% -15.7% 33280 ± 8% softirqs.CPU87.SCHED 135851 ± 5% -19.3% 109698 ± 2% softirqs.CPU87.TIMER 30947 ± 2% +20.1% 37173 ± 9% softirqs.CPU9.SCHED 133360 ± 3% -14.3% 114339 ± 3% softirqs.CPU9.TIMER 4299 ±103% -66.6% 1438 ± 9% softirqs.NET_RX 1788927 ± 16% -15.2% 1516780 ± 4% softirqs.RCU 11741892 -15.5% 9927286 softirqs.TIMER Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Rong Chen