This document was uploaded by user and they confirmed that they have the permission to share
it. If you are author or own the copyright of this book, please report to us by using this DMCA
report form. Report DMCA
Overview
Download & View Memory Tuning In Java as PDF for free.
Objectives • Have an insight into key aspects of JVM Runtime Performance and understanding of the means and methods to tune the JVM for optimal performance. • At the end of this session you should have: High level overview of the Java Virtual Machine (JVM) and its key components. Understanding of different Garbage Collection schemas in the Sovereign & J9 Virtual Machines and their impact on JVM Runtime performance.
Knowledge about using verbosegc outputs effectively to improve application response times.
Introduction to Shared Classes Technology and performance gains. Familiarity with debugging and profiling tools available for JVM and their effective usage.
Agenda • Overview of the Java Virtual Machine and its key components • Garbage Collection Basics (Sovereign VM – 142 JDK) • Profiling Garbage Collection: Verbosegc Outputs • Garbage Collection Policies in Java 5.0 • Debugging and Analysis tools for Garbage Collection • Introduction to Shared Classes Technology • Real Time Java – A brief overview • Q&A
Garbage Collection Overview ¾ Garbage Collection (GC) The main cause of memory–related performance bottlenecks in Java.
¾ Two things to look at in GC: frequency and duration Frequency depends on the heap size and allocation rate Duration depends on the heap size and number of objects in the heap
¾ GC algorithm Critical to understand how it works so that tuning is done more intelligently.
¾ How do you eliminate GC bottlenecks? Minimize the use of objects by following good programming practices Set your heap size properly, memory-tune your JVM
• “The purpose of Garbage Collection is to identify Java Heap storage which is no longer being used by the Java application and so is available for reuse by the JVM” •Key questions: •Performance and Scalability: How quickly can you find garbage? •Accuracy:
Garbage Collection: IBM Technology • Concurrent mark Most of the marking phase is done outside of ‘Stop the World’ when the ‘mutator’ threads are still active giving a significant improvement in pause time.
• Parallelizing the garbage Collection phases The Mark and sweep workload is distributed across available processors resulting in a significant improvement in pause times
• Adaptive sizing of thread local heaps Reduces the amount of Java Heap locking
• Incremental compaction The expense of compaction is distributed across GCs leading to a reduction in (an occasional) long pause time.
• Java 5 technologies Lazy sweep, parallel compaction, generational and accurate collection
Cache Allocation (for object allocations < 512 bytes), does not require Heap Lock. Each thread has local storage in the heap (TLH – Thread Local Heap) where the objects are allocated.
–
Heap Lock Allocation (Heap Allocation occurs when the allocation request is more than 512 bytes, requires Heap Lock. If size is less than 512 or enough space in the cache try cacheAlloc return if OK HEAP_LOCK do forever If there is a big enough chunk on freelist Take it goto Gotit else manageAllocFailure If any error goto GetOut End do Gotit: Initialize object Get out Heap_UNLOCK
All objects => 64K are termed “large” from the VM perspective
•
In practice, objects of 10MB+ in size are usually considered large
•
The Large Object Area is 5% of the active heap by default.
•
Any object is first tried to be allocated in the free list of the main heap – if there is not enough “contiguous” space in the main heap to satisfy the allocation request for object => 64K, then it is allocated in the Large Object Area (wilderness)
•
Objects < 64K can only be allocated in the main heap and never in the Large Object Area
Active heap
LOA
Xmx
Users can identify the Java stack of a thread making an allocation request of larger than the value specified with the environment variable ALLOCATION_THRESHOLD export ALLOCATION_THRESHOLD =5400 This will give java stacks for object allocations of created than 5400 bytes.
Users can specify the desired % of the Large Object Area using the Xloration option (where n determines the fraction of heap designated for LOA. (For example: Xloratio0.3 reserves 30% of the active heap for the Large Object Area)
Sub pools • Subpools provide an improved policy of object allocation and is available from JDK 1.4.1 releases only on AIX. – – – –
Improved time for allocating objects Avoid premature GCs due to allocation of large objects Improve MP scalability by reducing time under HEAP_LOCK Optimize TLH sizes and storage utilization
• The subpool algorithm uses multiple free lists rather than the single free list used by the default allocation scheme. • It tries to predict the size of future allocation requests based on earlier allocation requests. It recreates free lists at the end of each GC based on these predictions. • While allocating objects on the heap, free chunks are chosen using a ″best fit″ method, as against the ″first fit″ method used in other algorithms. • It is enabled by the –Xgcpolicy:subpool option. • Concurrent marking is disabled when subpool policy is used.
Garbage Collection Basics (142 JVM) • Garbage Collection is performed when there is: ¾ An allocation failure in the heap lock allocation ¾ Specific call to System.gc • Garbage Collection is Stop the World (All other application threads are suspended during GC) • Two main technologies used to remove garbage: ¾ Mark Sweep Collector ¾ Copy Collector GC occurs in the thread that handled the request
¾ Requested object allocation that caused allocation failure ¾ Programmatically requested GC • Thread must acquire certain locks required for GC ¾ Heap Lock ¾ Thread queue lock ¾ JIT lock
Together we achieve: Some parallel processing • GC Helper threads On a multiprocessor system with N CPUs, a JVM supporting parallel mode starts N-1 garbage collection helper threads at the time of initialization. These threads remain idle at all times when the application code is running; they are called into play only when garbage collection is active. For a particular phase, work is divided between the thread driving the garbage collection and the helper threads, making a total of N threads running in parallel on an N-CPU machine. The only way to disable the parallel mode is to use the -Xgcthreads parameter to change the number of garbage collection helper threads being started. • Parallel Mark The basic idea is to augment object marking through the addition of helper threads and a facility for sharing work between them. • Parallel BitWise Sweep Similar to parallel mark, uses same helper threads as parallel mark Improves sweep times by using available processors.
Concurrent Marking Designed to give reduced and consistent GC pause times as heap sizes increases. Concurrent aims to complete the marking just before the before the heap is full. In the concurrent phase, the Garbage Collector scans the roots by asking each thread to scan its own stack. These roots are then used to trace live objects concurrently. Tracing is done by a low-priority background thread and by each application thread when it does a heap lock allocation. Concurrent is enabled by the option: -Xgcpolicy:optavgpause
Incremental Compaction • Incremental compaction removes the dark matter from the heap and reduces pause times significantly • The fundamental idea behind incremental compaction is to split the heap up into sections and compact each section just as during a full compaction. • Incremental compaction was introduced in JDK 1.4.0; is enabled by default and triggered under particular conditions. (Called Reasons) • -Xpartialcompactgc • -Xnopartialcompactgc • -Xnocompactgc
Option to invoke incremental compaction in every GC cycle. Option to disable incremental compcation. Option to disable full compcation
The heap is divided into regions The regions are further divided into sections Each section is handled by one helper thread A region is divided into (number of helper threads +1) or 8 sections (whichever is less) The whole heap is covered in a few GC cycles.
Explicit Garbage Collection Garbage collector is called only upon two conditions: When an allocation failure occurs GC explicitly called using System.gc Don’t call System.gc() at all. It hurts more often than it helps. GC knows when it should run The temptation to scatter System.gc() calls here, there, and everywhere is enormous. It does not make a good idea. TRUST ME !!!!! Use –Xdisableexplicitgc to avoid running GC for each System.gc().
Profiling Garbage Collection: Verbosegc output The most indispensable tool for profiling GC activity is Verbosegc – from JVM runtime. Enabled using –verbosegc on the java command line. Verbosegc redirection -Xverbosegclog: <path to file> filename Verbosegc redirection to multiple files -Xverbosegclog:<path to file>filename#,X,Y (Redirect to Y files each containing X GC cycles)
When are GC messages printed out? • The first two lines are put out just before the beginning of STW phase of GC. • Rest of messages are printed out after the STW phase ends and threads are woken up. No messages are printed during GC. • Heap shrinkage messages are printed before STW messages, but shrinkage happens AFTER STW phase! • Heap expansion messages are correctly printed AFTER STW messages.
Things to look for in a verbosegc output • Was it an Allocation Failure GC? • What was the size of allocation request that caused AF? • What were the total and free heap sizes before GC? • What was the total pause time? • Where was maximum time spent in GC? • Are we doing a compaction in each GC cycle? • What actions were taken by GC? • Was GC able to meet allocation request in the end? • What was the free heap size after GC?
Using verbosegc to set the heap size • Use verbosegc to guess ideal size of heap, and then tune using –Xmx and –Xms. • Setting –Xms: Should be big enough to avoid AFs from the time the application starts to the time it becomes ‘ready’. (Should not be any bigger!)
• Setting –Xmx: In the normal load condition, free heap space after each GC should be > minf (Default is 30%). There should not be any OutOfMemory errors. In heaviest load condition, if free heap space after each GC is > maxf (Default is 70%), heap size is too big.
Effect of wrong –Xms & -Xmx settings Too small heap = Too frequent GC. Too big heap = Too much GC pause time. (Irrespective of amount of physical memory on the system) Heap size > physical memory size = paging/swapping = bad for your application. It is desirable to have the Xms much less than Xmx if you are encountering fragmentation issues. This forces class allocations, thread and persistent objects to be allocated at the bottom of the heap.
What about Xms=Xmx? It means no heap expansion or shrinkage will ever occur. Not normally recommended. It may be good for a few apps which require constant high heap storage space. Hurts apps which show a varying load. May lead to more fragmentation
Mark Stack Overflow (MSO) • Verbosegc will contain the line: • Is bad for performance. • Caused by too many objects on the heap, especially deeply nested objects. • Processing MSOs is expensive • No solution is a silver bullet. Things to try: ¾ Decrease the heap size! ¾ Use concurrent mark (-Xgcpolicy:optavgpause) ¾ Re-design the application. ¾ Increase GC helper threads (-Xgcthreads)
High pause times and system activity In the event of pause times being usually acceptable with the exception of a few "abnormally high" spikes - we are likely to infer that the deviation was a result of some system level activity (heavy paging for ex) outside of the Java process. Consideration: How many clock ticks our process actually spent executing instructions, not time spent waiting for I/O or time spent waiting for a CPU to become available for the process to run on?
Java Tiger: JSE 5.0 ¾ 5.0 uses a completely new memory management framework ¾ No pinned/dosed objects Stack Maps used to provide a level of indirection between references and heap 5.0 VM never pins arrays, it always makes a copy for the JNI code ¾ The GC is Type Accurate ¾ New efficient parallel compactor ¾ -Xgcpolicy:optavgpause includes concurrent sweeping (as well as marking)
Garbage collection policies in IBM Java 5.0 • Four policies available: Optthruput (default) ¾ Mark-sweep algorithm ¾ Fastest for many workloads
optavgpause ¾ Concurrent collection and concurrent sweep ¾ Small mean pause ¾ Throughput impact
gencon ¾ The Generational Hypothesis ¾ Fastest for transactional workloads ¾ Combines low pause times and high throughput
subpools ¾ Mark-sweep based, but with multiple freelists ¾ Avoids allocation contention on many-processor machines ¾ Available on all PPC and S390 platforms
Why different GC Policies? • Availability of different GC policies gives you increased capabilities. • Best choice depends upon application behaviour and workloads. • Think about throughput, response times & pause times.
Policy Throughput is the amount of data processed by the application Response time is the latency of the application – how quickly it answers incoming requests Pause time is the amount of time the garbage collector has stopped threads while collecting the heap.
Optimize for throughput
Option ( -Xgcpolicy ) Optthruput (Default)
Description It is typically used for applications where raw throughput is more important than short GC pauses. The application is stopped each time that garbage is collected.
Optimize for pause times
Optavgpause
Trades high throughput for shorter GC pauses by performing some of the garbage collection concurrently. The application is paused for shorter times.
Generational Concurrent
gencon
Handles short lived objects differently than the longer lived.
Supool
subpool
Uses same algorithm similar to the default policy but employs allocation strategy suitable for SMPs.
Runtime Performance Tuning What policy should I choose for my J9 VM
Policy
Considerations
optthruput
I want my application to run to completion as quickly as possible.
optavgpause
• My application requires good response times to unpredictable events. • A degradation in performance is acceptable as long as GC pause times are reduced. • My application is running very large java heaps. • My application is a GUI application and user response times are important.
gencon
• My application allocates many short lived objects • The java heap space is fragmented • My application is transactions based
Nursery/Young Generation Nursery/Young Generation Survivor Space Allocate Space
Allocate Survivor Space Space
• Nursery is split into two spaces (semi-spaces) Only one contains live objects and is available for allocation Minor collections (Scavenges) move objects between spaces Role of spaces is reversed
Diagnostic tool for garbage collector • Diagnostic tool for optimizing parameters affecting the garbage collector while using IBM JVM • Reads the “verbosegc” output and produces textual and graphical visualizations and related statistics – – – – –
Frequency of garbage collection cycles Time spent in different phases of garbage collection Quantities of heap memories involved in the process Characteristics of allocation failures Mark Stack Overflows
• Built in parsers for JVM 1.5, 1.4.2 , 1.3.1 & 1.2.2 • Prerequisite – JFreeChart libraries (freely downloadable) • http://www.alphaworks.ibm.com/tech/gcdiag
Starting the tool and selecting verbosegc input file and appropriate JVM parser
Extract GCCollector.zip Place jfreeChart-0.9.21.jar and jcommon-0.9.6.jar in the lib directory of the GCCollector folder Execute GCCollector.bat (which will spawn a GUI) Select verbosegc file for analysis Select appropriate JVM parser
Extensible Verbose Tool Kit Analyzing your verbose GC output
• EVTK: Verbose GC visualizer and analyzer
Available through ISA ¾ IBM Support Assistant v3.0.2 https://www14.software.ibm.com/webapp/iwm/web/preLogin.do?source=isa
Reduces barrier to understanding verbose output ¾Visualize GC data to track trends and relationships ¾ Analyze results and provide general feedback ¾ Extend to consume output related to application ¾ Build plug-able analyzers for specific application needs
Extensible Verbose Toolkit overview • The Extensible Verbose Toolkit (EVTK) is a visualizer for verbose garbage collection output The tool parses and plots verbose GC output and garbage collection traces (-Xtgc output)
• The tooling framework is extensible, and will be expanded over time to include visualization for other collections of data • The EVTK provides
Raw view of data Line plots to visualize a variety of GC data characteristics Tabulated reports with heap occupancy recommendations View of multiple datasets on a single set of axes Ability to save data as an image (jpeg) or comma separated file (csv)
The VGC Data menu allows you to choose what data to display
Use File > Add to add multiple input files to a single data set for comparison and aggregated display Right-click on the plot and use the context menu to export data
Reports and recommendations • Report contents can be configured using VGC menu options Occupancy recommendations tell you how to adjust heap size for better performance Summary information is generated for each input in the dataset Graphs included for all GC display data • Can export as HTML by rightclicking and using the context menu
The Report tab contains the report for the current dataset
Types of graphs • The EVTK has built-in support for over forty different types of graphs These are configured in the VGC Data menu Options vary depending on the current dataset and the parsers and postprocessors that are enabled • Some of the VGC graph types are: • Used total heap • Pause times (mark-sweep-compact collections) • Pause times (totals, including exclusive access) • Compact times • Weak references cleared • Soft references cleared
• • • • • • •
Free tenured heap (after collection) Tenured heap size Tenure age Free LOA (after collection) Free SOA (after collection) Total LOA Total SOA
• Note: Different graph types and a different menu are available for TGC output
Heap usage and occupancy recommendation The summary report shows that mean heap occupancy is 98% and that the application is spending over a third of its time doing garbage collection
Java 5 Shared Classes • Available on all platforms. • Feature enabled using the –Xshareclasses flag. •
Static class data caches in shared memory Shared between all IBM Java VMs All application and bootstrap classes shared Cache persisted beyond lifetime of any JVM, but lost on shutdown/reboot
•
Provides savings in footprint and start up times.
• Target: Server environments where multiple JVMs exist on the same box. • Multiple sharing strategies Standard Classloaders (including Application Classloader) exploit this feature when enabled. API to extend Customer Classloaders available.
Real-time = predictability of performance Hard - Violation of timing constraints are hard failures Soft - Timing constraints are simply performance goals
•
Constraints vary in magnitude (microseconds to seconds)
•
Consequences of missing a timing constraint: from service level agreement miss (stock trading) to life in jeopardy (airplanes)
•
Real-fast is not real-time, but Real-slow is not real-good
•
Need a balance between predictability and throughput
WRT is a Java runtime providing highly predictable operation Real-time garbage collection (Metronome) Static and dynamic compilation Full support for RTSJ (JSR 1) Java SE 5.0 compliant Built and rigorously tested on a RT Linux OS with IBM Opteron Blades
Visual Performance Analyser • Eclipse based performance visualization kit – Profile Analyzer – Code Analyzer – Pipeline Analyzer • Profiler Analyzer provides a powerful set of graphical and text-based views that allow users to narrow down performance problems to a particular process, thread, module, symbol, offset, instruction or source line. – Supports time based system profiles • Code Analyzer examines executable files and displays detailed information about functions, basic blocks and assembly instructions. • Pipeline Analyzer displays the pipeline execution of instruction traces. • http://www.alphaworks.ibm.com/tech/vpa
Summary • Take Aways from the Session: High level overview of the Java Virtual Machine (JVM) and its key components. Understanding of different Garbage Collection schemas in the Sovereign & J9 Virtual Machines and their impact on JVM Runtime performance. Knowledge about using verbosegc outputs effectively to improve application response times. Introduction to Shared Classes Technology and performance gains. Familiarity with debugging and profiling tools available for JVM and their effective usage.