comparison src/share/vm/runtime/globals.hpp @ 14702:d4dd5204c0aa

Merge
author ehelin
date Fri, 14 Mar 2014 13:27:18 +0100
parents 480b0109db65
children b51e29501f30
comparison
equal deleted inserted replaced
14701:d3ef21095134 14702:d4dd5204c0aa
533 "Use lwsync instruction if true, else use slower sync") \ 533 "Use lwsync instruction if true, else use slower sync") \
534 \ 534 \
535 develop(bool, CleanChunkPoolAsync, falseInEmbedded, \ 535 develop(bool, CleanChunkPoolAsync, falseInEmbedded, \
536 "Clean the chunk pool asynchronously") \ 536 "Clean the chunk pool asynchronously") \
537 \ 537 \
538 experimental(bool, AlwaysSafeConstructors, false, \
539 "Force safe construction, as if all fields are final.") \
540 \
538 /* Temporary: See 6948537 */ \ 541 /* Temporary: See 6948537 */ \
539 experimental(bool, UseMemSetInBOT, true, \ 542 experimental(bool, UseMemSetInBOT, true, \
540 "(Unstable) uses memset in BOT updates in GC code") \ 543 "(Unstable) uses memset in BOT updates in GC code") \
541 \ 544 \
542 diagnostic(bool, UnlockDiagnosticVMOptions, trueInDebug, \ 545 diagnostic(bool, UnlockDiagnosticVMOptions, trueInDebug, \
809 "Trace calls to C heap malloc/free allocation") \ 812 "Trace calls to C heap malloc/free allocation") \
810 \ 813 \
811 product(bool, PrintOopAddress, false, \ 814 product(bool, PrintOopAddress, false, \
812 "Always print the location of the oop") \ 815 "Always print the location of the oop") \
813 \ 816 \
814 notproduct(bool, VerifyCodeCacheOften, false, \ 817 notproduct(bool, VerifyCodeCache, false, \
815 "Verify compiled-code cache often") \ 818 "Verify code cache on memory allocation/deallocation") \
816 \ 819 \
817 develop(bool, ZapDeadCompiledLocals, false, \ 820 develop(bool, ZapDeadCompiledLocals, false, \
818 "Zap dead locals in compiler frames") \ 821 "Zap dead locals in compiler frames") \
819 \ 822 \
820 notproduct(bool, ZapDeadLocalsOld, false, \ 823 notproduct(bool, ZapDeadLocalsOld, false, \
2982 \ 2985 \
2983 product(intx, MaxRecursiveInlineLevel, 1, \ 2986 product(intx, MaxRecursiveInlineLevel, 1, \
2984 "maximum number of nested recursive calls that are inlined") \ 2987 "maximum number of nested recursive calls that are inlined") \
2985 \ 2988 \
2986 develop(intx, MaxForceInlineLevel, 100, \ 2989 develop(intx, MaxForceInlineLevel, 100, \
2987 "maximum number of nested @ForceInline calls that are inlined") \ 2990 "maximum number of nested calls that are forced for inlining " \
2991 "(using CompilerOracle or marked w/ @ForceInline)") \
2988 \ 2992 \
2989 product_pd(intx, InlineSmallCode, \ 2993 product_pd(intx, InlineSmallCode, \
2990 "Only inline already compiled methods if their code size is " \ 2994 "Only inline already compiled methods if their code size is " \
2991 "less than this") \ 2995 "less than this") \
2992 \ 2996 \
3290 "Thread safety margin is used on fixed-stack LinuxThreads (on " \ 3294 "Thread safety margin is used on fixed-stack LinuxThreads (on " \
3291 "Linux/x86 only) to prevent heap-stack collision. Set to 0 to " \ 3295 "Linux/x86 only) to prevent heap-stack collision. Set to 0 to " \
3292 "disable this feature") \ 3296 "disable this feature") \
3293 \ 3297 \
3294 /* code cache parameters */ \ 3298 /* code cache parameters */ \
3295 /* ppc64 has large code-entry alignment. */ \ 3299 /* ppc64/tiered compilation has large code-entry alignment. */ \
3296 develop(uintx, CodeCacheSegmentSize, 64 PPC64_ONLY(+64), \ 3300 develop(uintx, CodeCacheSegmentSize, 64 PPC64_ONLY(+64) NOT_PPC64(TIERED_ONLY(+64)),\
3297 "Code cache segment size (in bytes) - smallest unit of " \ 3301 "Code cache segment size (in bytes) - smallest unit of " \
3298 "allocation") \ 3302 "allocation") \
3299 \ 3303 \
3300 develop_pd(intx, CodeEntryAlignment, \ 3304 develop_pd(intx, CodeEntryAlignment, \
3301 "Code entry alignment for generated code (in bytes)") \ 3305 "Code entry alignment for generated code (in bytes)") \
3793 "show method handle implementation frames (usually hidden)") \ 3797 "show method handle implementation frames (usually hidden)") \
3794 \ 3798 \
3795 experimental(bool, TrustFinalNonStaticFields, false, \ 3799 experimental(bool, TrustFinalNonStaticFields, false, \
3796 "trust final non-static declarations for constant folding") \ 3800 "trust final non-static declarations for constant folding") \
3797 \ 3801 \
3798 experimental(bool, FoldStableValues, false, \ 3802 diagnostic(bool, FoldStableValues, true, \
3799 "Private flag to control optimizations for stable variables") \ 3803 "Optimize loads from stable fields (marked w/ @Stable)") \
3800 \ 3804 \
3801 develop(bool, TraceInvokeDynamic, false, \ 3805 develop(bool, TraceInvokeDynamic, false, \
3802 "trace internal invoke dynamic operations") \ 3806 "trace internal invoke dynamic operations") \
3803 \ 3807 \
3804 diagnostic(bool, PauseAtStartup, false, \ 3808 diagnostic(bool, PauseAtStartup, false, \
3861 \ 3865 \
3862 experimental(uintx, ArrayAllocatorMallocLimit, \ 3866 experimental(uintx, ArrayAllocatorMallocLimit, \
3863 SOLARIS_ONLY(64*K) NOT_SOLARIS(max_uintx), \ 3867 SOLARIS_ONLY(64*K) NOT_SOLARIS(max_uintx), \
3864 "Allocation less than this value will be allocated " \ 3868 "Allocation less than this value will be allocated " \
3865 "using malloc. Larger allocations will use mmap.") \ 3869 "using malloc. Larger allocations will use mmap.") \
3870 \
3871 experimental(bool, AlwaysAtomicAccesses, false, \
3872 "Accesses to all variables should always be atomic") \
3866 \ 3873 \
3867 product(bool, EnableTracing, false, \ 3874 product(bool, EnableTracing, false, \
3868 "Enable event-based tracing") \ 3875 "Enable event-based tracing") \
3869 \ 3876 \
3870 product(bool, UseLockedTracing, false, \ 3877 product(bool, UseLockedTracing, false, \