OS Images |
os_Image_1(1)
|
Hardware Description |
hw_1
|
Number of Systems |
1
|
SW Environment |
Non-virtual
|
Tuning |
- Workload Profile=High Performance Compute(HPC)
- Thermal Configuration=Maximum Cooling
- Determinism Control=Manual
- Performance Determinism=Power Deterministic
- Memory Patrol Scrubbing=Disabled
- Numa Memory Domains Per Socket(NPS)=Four Memory Domains Per Socket
- Last-Level Cache(LLC) As NUMA Node=Enabled
- L1 Stream HW Prefetcher=Disabled
- L2 Stream HW Prefetcher=Disabled
- Minimum Processor Idle Power Core C-State=No C-states
- xGMI Link Bandwidth=32Gbps
- Package Power Limit=400
|
Notes |
None
|
|
JVM Instances |
jvm_Ctr_1(1), jvm_Backend_1(24), jvm_TxInjector_1(24)
|
OS Image Description |
os_1
|
Tuning |
- ulimit -n 1024000
- UserTasksMax=970000
- DefaultTasksMax=970000
- tuned-adm profile throughput-performance
- write_schedstat rt_runtime_us 960000
- write_schedstat latency_ns 20000000
- write_schedstat migration_cost_ns 40000
- write_schedstat min_granularity_ns 810000000
- write_schedstat wakeup_granularity_ns 200000000
- write_schedstat nr_migrate 9000
- echo 10000 > /proc/sys/vm/dirty_expire_centisecs
- echo 1500 > /proc/sys/vm/dirty_writeback_centisecs
- echo 40 > /proc/sys/vm/dirty_ratio
- echo 10 > /proc/sys/vm/dirty_background_ratio
- echo 10 > /proc/sys/vm/swappiness
- echo 0 > /proc/sys/kernel/numa_balancing
- echo 0 > /proc/sys/vm/numa_stat
- echo always > /sys/kernel/mm/transparent_hugepage/enabled
- echo always > /sys/kernel/mm/transparent_hugepage/defrag
|
Notes |
None
|
Parts of Benchmark |
Controller
|
JVM Instance Description |
jvm_1
|
Command Line |
-Xms8g -Xmx8g -Xmn6g -XX:+UseParallelGC -XX:ParallelGCThreads=1 -XX:CICompilerCount=2
|
Tuning |
Used numactl to interleave memory on all CPUs
|
Notes |
None
|
Parts of Benchmark |
Backend
|
JVM Instance Description |
jvm_1
|
Command Line |
-Xms31g -Xmx31g -Xmn29g -XX:AllocatePrefetchInstr=2 -XX:+UseParallelGC -XX:ParallelGCThreads=16 -XX:LargePageSizeInBytes=2m -XX:-UseAdaptiveSizePolicy -XX:+AlwaysPreTouch -XX:+UseLargePages -XX:SurvivorRatio=12 -XX:TargetSurvivorRatio=95 -XX:MaxTenuringThreshold=15 -XX:InlineSmallCode=11k -XX:MaxGCPauseMillis=100 -XX:LoopUnrollLimit=200 -XX:+UseTransparentHugePages -XX:TLABAllocationWeight=2 -XX:ThreadStackSize=140 -XX:CompileThresholdScaling=120 -XX:CICompilerCount=4 -XX:AutoBoxCacheMax=32 -XX:OnStackReplacePercentage=100 -XX:TLABSize=1m -XX:MinTLABSize=1m -XX:-ResizeTLAB -XX:TLABWasteTargetPercent=1 -XX:TLABWasteIncrement=1 -XX:YoungPLABSize=1m -XX:OldPLABSize=1m
|
Tuning |
Used numactl to affinitize each Backend JVM to physical cores in a NUMA node. - numactl --physcpubind=0-7,192-199 --localalloc
- numactl --physcpubind=8-15,200-207 --localalloc
- numactl --physcpubind=16-23,208-215 --localalloc
- numactl --physcpubind=24-31,216-223 --localalloc
- numactl --physcpubind=32-39,224-231 --localalloc
- numactl --physcpubind=40-47,232-239 --localalloc
- numactl --physcpubind=48-55,240-247 --localalloc
- numactl --physcpubind=56-63,248-255 --localalloc
- numactl --physcpubind=64-71,256-263 --localalloc
- numactl --physcpubind=72-79,264-271 --localalloc
- numactl --physcpubind=80-87,272-279 --localalloc
- numactl --physcpubind=88-95,280-287 --localalloc
- numactl --physcpubind=96-103,288-295 --localalloc
- numactl --physcpubind=104-111,296-303 --localalloc
- numactl --physcpubind=112-119,304-311 --localalloc
- numactl --physcpubind=120-127,312-319 --localalloc
- numactl --physcpubind=128-135,320-327 --localalloc
- numactl --physcpubind=136-143,328-335 --localalloc
- numactl --physcpubind=144-151,336-343 --localalloc
- numactl --physcpubind=152-159,344-351 --localalloc
- numactl --physcpubind=160-167,352-359 --localalloc
- numactl --physcpubind=168-175,360-367 --localalloc
- numactl --physcpubind=176-183,368-375 --localalloc
- numactl --physcpubind=184-191,376-383 --localalloc
|
Notes |
None
|
Parts of Benchmark |
TxInjector
|
JVM Instance Description |
jvm_1
|
Command Line |
-Xms8g -Xmx8g -Xmn6g -XX:+UseParallelGC -XX:ParallelGCThreads=1 -XX:CICompilerCount=2
|
Tuning |
Used numactl to affinitize each Transaction Injector JVM to physical cores in a NUMA node. - numactl --physcpubind=0-7,192-199 --localalloc
- numactl --physcpubind=8-15,200-207 --localalloc
- numactl --physcpubind=16-23,208-215 --localalloc
- numactl --physcpubind=24-31,216-223 --localalloc
- numactl --physcpubind=32-39,224-231 --localalloc
- numactl --physcpubind=40-47,232-239 --localalloc
- numactl --physcpubind=48-55,240-247 --localalloc
- numactl --physcpubind=56-63,248-255 --localalloc
- numactl --physcpubind=64-71,256-263 --localalloc
- numactl --physcpubind=72-79,264-271 --localalloc
- numactl --physcpubind=80-87,272-279 --localalloc
- numactl --physcpubind=88-95,280-287 --localalloc
- numactl --physcpubind=96-103,288-295 --localalloc
- numactl --physcpubind=104-111,296-303 --localalloc
- numactl --physcpubind=112-119,304-311 --localalloc
- numactl --physcpubind=120-127,312-319 --localalloc
- numactl --physcpubind=128-135,320-327 --localalloc
- numactl --physcpubind=136-143,328-335 --localalloc
- numactl --physcpubind=144-151,336-343 --localalloc
- numactl --physcpubind=152-159,344-351 --localalloc
- numactl --physcpubind=160-167,352-359 --localalloc
- numactl --physcpubind=168-175,360-367 --localalloc
- numactl --physcpubind=176-183,368-375 --localalloc
- numactl --physcpubind=184-191,376-383 --localalloc
|
Notes |
None
|
|