OS Images |
os_Image_1(1)
|
Hardware Description |
hw_1
|
Number of Systems |
1
|
SW Environment |
Non-virtual
|
Tuning |
- Dram Refresh Delay - Performance
- DIMM Self Healing (Post Package Repair) on Uncorrectable Memory Error - Disabled
- Correctable Error Logging - Disabled
- Virtualization Technology - Disabled
- L2 Stream HW Prefetcher - Disabled
- NUMA Nodes Per Socket - 4
- L3 cache as NUMA Domain- Enabled
- TDP Control- Manual
- Customized cTDP - 500
- Customized PPT - 500
- System Profile - Custom
- CPU Power Management - Maximum Performance
- C-States - Disabled
- Determinism Control- Manual
- Determinism Slider - Power Determinism
|
Notes |
None
|
|
JVM Instances |
jvm_Ctr_1(1), jvm_Backend_1(24), jvm_TxInjector_1(24)
|
OS Image Description |
os_1
|
Tuning |
- echo 960000 > /proc/sys/kernel/sched_rt_runtime_us
- echo 999000000 > /proc/sys/kernel/sched_latency_ns
- echo 40000 > /proc/sys/kernel/sched_migration_cost_ns
- echo 999000000 > /proc/sys/kernel/sched_min_granularity_ns
- echo 999000000 > /proc/sys/kernel/sched_wakeup_granularity_ns
- echo 9000 > /proc/sys/kernel/sched_nr_migrate
- echo 10000 > /proc/sys/vm/dirty_expire_centisecs
- echo 1500 > /proc/sys/vm/dirty_writeback_centisecs
- echo 40 > /proc/sys/vm/dirty_ratio
- echo 10 > /proc/sys/vm/dirty_background_ratio
- echo 10 > /proc/sys/vm/swappiness
- echo 0 > /proc/sys/vm/numa_stat
- echo 0 > /proc/sys/kernel/numa_balancing
- echo always > /sys/kernel/mm/transparent_hugepage/enabled
- echo always > /sys/kernel/mm/transparent_hugepage/defrag
|
Notes |
None
|
Parts of Benchmark |
Controller
|
JVM Instance Description |
jvm_1
|
Command Line |
-Xms3g -Xmx3g -Xmn2g -XX:+UseParallelGC -XX:ParallelGCThreads=1 -XX:CICompilerCount=2
|
Tuning |
Used numactl to interleave memory on all CPUs
|
Notes |
None
|
Parts of Benchmark |
Backend
|
JVM Instance Description |
jvm_1
|
Command Line |
-Xms31g -Xmx31g -Xmn29g -XX:AllocatePrefetchInstr=2 -XX:+UseParallelGC -XX:ParallelGCThreads=32 -XX:LargePageSizeInBytes=2m -XX:-UseAdaptiveSizePolicy -XX:+AlwaysPreTouch -XX:+UseLargePages -XX:SurvivorRatio=12 -XX:TargetSurvivorRatio=95 -XX:MaxTenuringThreshold=15 -XX:InlineSmallCode=11k -XX:MaxGCPauseMillis=100 -XX:LoopUnrollLimit=200 -XX:+UseTransparentHugePages -XX:TLABAllocationWeight=2 -XX:ThreadStackSize=140 -XX:CompileThresholdScaling=120 -XX:CICompilerCount=4 -XX:AutoBoxCacheMax=32 -XX:OnStackReplacePercentage=100 -XX:TLABSize=1m -XX:MinTLABSize=1m -XX:-ResizeTLAB -XX:TLABWasteTargetPercent=1 -XX:TLABWasteIncrement=1 -XX:YoungPLABSize=1m -XX:OldPLABSize=1m
|
Tuning |
Used numactl to affinitize each Backend JVM to 16 Cores / 32 Threads: - Group1: numactl --physcpubind=0-15,384-399 --localalloc
- Group2: numactl --physcpubind=16-31,400-415 --localalloc
- Group3: numactl --physcpubind=32-47,416-431 --localalloc
- Group4: numactl --physcpubind=48-63,432-447 --localalloc
- Group5: numactl --physcpubind=64-79,448-463 --localalloc
- Group6: numactl --physcpubind=80-95,464-479 --localalloc
- Group7: numactl --physcpubind=96-111,480-495 --localalloc
- Group8: numactl --physcpubind=112-127,496-511 --localalloc
- Group9: numactl --physcpubind=128-143,512-527 --localalloc
- Group10: numactl --physcpubind=144-159,528-543 --localalloc
- Group11: numactl --physcpubind=160-175,544-559 --localalloc
- Group12: numactl --physcpubind=176-191,560-575 --localalloc
- Group13: numactl --physcpubind=192-207,576-591 --localalloc
- Group14: numactl --physcpubind=208-223,592-607 --localalloc
- Group15: numactl --physcpubind=224-239,608-623 --localalloc
- Group16: numactl --physcpubind=240-255,624-639 --localalloc
- Group17: numactl --physcpubind=256-271,640-655 --localalloc
- Group18: numactl --physcpubind=272-287,656-671 --localalloc
- Group19: numactl --physcpubind=288-303,672-687 --localalloc
- Group20: numactl --physcpubind=304-319,688-703 --localalloc
- Group21: numactl --physcpubind=320-335,704-719 --localalloc
- Group22: numactl --physcpubind=336-351,720-735 --localalloc
- Group23: numactl --physcpubind=352-367,736-751 --localalloc
- Group24: numactl --physcpubind=368-383,752-767 --localalloc
|
Notes |
None
|
Parts of Benchmark |
TxInjector
|
JVM Instance Description |
jvm_1
|
Command Line |
-Xms3g -Xmx3g -Xmn2g -XX:+UseParallelGC -XX:ParallelGCThreads=1 -XX:CICompilerCount=2
|
Tuning |
Used numactl to affinitize each TxInjector JVM to 16 Cores / 32 Threads : - Group1: numactl --physcpubind=0-15,384-399 --localalloc
- Group2: numactl --physcpubind=16-31,400-415 --localalloc
- Group3: numactl --physcpubind=32-47,416-431 --localalloc
- Group4: numactl --physcpubind=48-63,432-447 --localalloc
- Group5: numactl --physcpubind=64-79,448-463 --localalloc
- Group6: numactl --physcpubind=80-95,464-479 --localalloc
- Group7: numactl --physcpubind=96-111,480-495 --localalloc
- Group8: numactl --physcpubind=112-127,496-511 --localalloc
- Group9: numactl --physcpubind=128-143,512-527 --localalloc
- Group10: numactl --physcpubind=144-159,528-543 --localalloc
- Group11: numactl --physcpubind=160-175,544-559 --localalloc
- Group12: numactl --physcpubind=176-191,560-575 --localalloc
- Group13: numactl --physcpubind=192-207,576-591 --localalloc
- Group14: numactl --physcpubind=208-223,592-607 --localalloc
- Group15: numactl --physcpubind=224-239,608-623 --localalloc
- Group16: numactl --physcpubind=240-255,624-639 --localalloc
- Group17: numactl --physcpubind=256-271,640-655 --localalloc
- Group18: numactl --physcpubind=272-287,656-671 --localalloc
- Group19: numactl --physcpubind=288-303,672-687 --localalloc
- Group20: numactl --physcpubind=304-319,688-703 --localalloc
- Group21: numactl --physcpubind=320-335,704-719 --localalloc
- Group22: numactl --physcpubind=336-351,720-735 --localalloc
- Group23: numactl --physcpubind=352-367,736-751 --localalloc
- Group24: numactl --physcpubind=368-383,752-767 --localalloc
|
Notes |
None
|
|