Table of Contents | ||
---|---|---|
|
Overview
- These tests are run to investigate the performance difference of mod-data-export with logLevel=info vs logLevel=warn for Juniper release. This testing was part of
where we observed that mod-data-export is continuously writing a lot of data to log. For 100K DE job, mod-data-export is writing 42Million records. This could also result in a crash if there is not enough CPU, memory allocated to mod-data-export.Jira Legacy server System JiraJIRA columnIds issuekey,summary,issuetype,created,updated,duedate,assignee,reporter,priority,status,resolution columns key,summary,type,created,updated,due,assignee,reporter,priority,status,resolution serverId 01505d01-b853-3c2e-90f1-ee9b165564fc key MDEXP-394 - In mod-data-export log4j2.properties file was modified with rootLogger.level = warn, logger.netty.level = warn, status = warn
...
- 8 million inventory records
- 74 FOLIO back-end modules deployed in 144 ECS services
- 3 okapi ECS services
- 12 m5.large EC2 instances
- 1 writer db.r6g.xlarge 1 reader db.r6g.xlarge AWS RDS instance
- INFO logging level / WARN logging level
- mod-data-export Soft memory limit - 360 MB Hard memory limit - 512 MB
- CPU - 128
High-Level Summary
- With WARN level logging, 9% improvement in memory utilization
- With WARN level logging, 70% improvement in CPU utilization
- With WARN level logging, No bumps in the memory were observed. Memory utilization stays stable for multiple Data Export job runs.
- Improvement by 42Million records for WARN. 42 Million fewer records were written
- With WARN level logging, there is improvement in JVM's Garbage collection
Test Runs
1 user - INFO level logging vs WARN level logging
...
INFO - ~42.5 Million records
WARN - ~120K records
Service Memory Utilization
For mod-data-export, 9% improvement in memory for WARN level logging
Memory Utilization INFO vs WARN
CPU Utilization
For mod-data-export, 70% improvement in CPU utilization for WARN level logging
INFO level logging
WARN level logging
Heap dump analysis using JXRay
4.6% improvement in Garbage collection with WARN level logging. That means, mod-data-export is generating less Unreachable garbage objects. WARN level logging is effectively utilization 12 MB of JVM memory. While with INFO level logging, this 12 MB is wasted. This is only for 100K run. As we run the Data Export job continuously, this 12MB could be much larger.
...
Other than this, with WARN level logging, mod-data-export will create fewer Object instances in JVM. For detailed comparison see mod-data-export-with-info-logging.html vs mod-data-export-with-warn-logging.html
Investigate GC with default info level logging
How much metaspace size should we allocate to mod-data-export?
Around 20%-25% of Xmx. Currently, mod-data-export is running with Xmx=388Mb and XX:MetaspaceSize=88Mb which is 22%.
We can check metaspace capacity with jstat
jstat -gcmetacapacity (PID)
For example:
root@2335b20abe22:/usr/ms# jstat -gcmetacapacity 60
MCMN MCMX MC CCSMN CCSMX CCSC YGC FGC FGCT CGC CGCT GCT
0.0 143360.0 70272.0 0.0 81920.0 9344.0 20690 451 43.211 - - 143.184
MC=70272Kb means currently, mod-data-export is using ~70MB of metspacesize
Checked how much percent of metaspacesize is used by mod-data-export
root@2335b20abe22:/usr/ms# jstat -gcutil 60 | awk '{print($5)}'
97.00
M=97.00 means 97% of metaspace size is used. Metaspace size allocated soft and hard is 88MB.
Investigated GC algorithm by using ShenandoahGC. But found out that JVM with OpenJDK package does not support it. By default it uses G1
Investigated another GC algorithm by using ZGC. After applying in the task definition, mod-data-export did become stable but it crashes after updating job profile in UI.
References
http://www.mastertheboss.com/java/solving-java-lang-outofmemoryerror-metaspace-error/
https://wiki.openjdk.java.net/display/zgc/Main#Main-JDK12
https://hub.packtpub.com/getting-started-with-z-garbage-collectorzgc-in-java-11-tutorial/