3763 #endif // COMPILER1 3764 ) { 3765 if (!FLAG_IS_DEFAULT(UseBiasedLocking) && UseBiasedLocking) { 3766 // flag set to true on command line; warn the user that they 3767 // can't enable biased locking here 3768 warning("Biased Locking is not supported with locking debug flags" 3769 "; ignoring UseBiasedLocking flag." ); 3770 } 3771 UseBiasedLocking = false; 3772 } 3773 3774 #ifdef ZERO 3775 // Clear flags not supported on zero. 3776 FLAG_SET_DEFAULT(ProfileInterpreter, false); 3777 FLAG_SET_DEFAULT(UseBiasedLocking, false); 3778 LP64_ONLY(FLAG_SET_DEFAULT(UseCompressedOops, false)); 3779 LP64_ONLY(FLAG_SET_DEFAULT(UseCompressedClassPointers, false)); 3780 #endif // CC_INTERP 3781 3782 #ifdef COMPILER2 3783 if (!UseBiasedLocking || EmitSync != 0) { 3784 UseOptoBiasInlining = false; 3785 } 3786 if (!EliminateLocks) { 3787 EliminateNestedLocks = false; 3788 } 3789 if (!Inline) { 3790 IncrementalInline = false; 3791 } 3792 #ifndef PRODUCT 3793 if (!IncrementalInline) { 3794 AlwaysIncrementalInline = false; 3795 } 3796 #endif 3797 if (IncrementalInline && FLAG_IS_DEFAULT(MaxNodeLimit)) { 3798 // incremental inlining: bump MaxNodeLimit 3799 FLAG_SET_DEFAULT(MaxNodeLimit, (intx)75000); 3800 } 3801 if (!UseTypeSpeculation && FLAG_IS_DEFAULT(TypeProfileLevel)) { 3802 // nothing to use the profiling, turn if off 3803 FLAG_SET_DEFAULT(TypeProfileLevel, 0); 3804 } 3805 if (UseTypeSpeculation && FLAG_IS_DEFAULT(ReplaceInParentMaps)) { 3826 } 3827 3828 if (!LogVMOutput && FLAG_IS_DEFAULT(LogVMOutput)) { 3829 if (use_vm_log()) { 3830 LogVMOutput = true; 3831 } 3832 } 3833 #endif // PRODUCT 3834 3835 if (PrintCommandLineFlags) { 3836 CommandLineFlags::printSetFlags(tty); 3837 } 3838 3839 // Apply CPU specific policy for the BiasedLocking 3840 if (UseBiasedLocking) { 3841 if (!VM_Version::use_biased_locking() && 3842 !(FLAG_IS_CMDLINE(UseBiasedLocking))) { 3843 UseBiasedLocking = false; 3844 } 3845 } 3846 3847 return JNI_OK; 3848 } 3849 3850 jint Arguments::adjust_after_os() { 3851 if (UseNUMA) { 3852 if (UseParallelGC || UseParallelOldGC) { 3853 if (FLAG_IS_DEFAULT(MinHeapDeltaBytes)) { 3854 FLAG_SET_DEFAULT(MinHeapDeltaBytes, 64*M); 3855 } 3856 } 3857 // UseNUMAInterleaving is set to ON for all collectors and 3858 // platforms when UseNUMA is set to ON. NUMA-aware collectors 3859 // such as the parallel collector for Linux and Solaris will 3860 // interleave old gen and survivor spaces on top of NUMA 3861 // allocation policy for the eden space. 3862 // Non NUMA-aware collectors such as CMS, G1 and Serial-GC on 3863 // all platforms and ParallelGC on Windows will interleave all 3864 // of the heap spaces across NUMA nodes. 3865 if (FLAG_IS_DEFAULT(UseNUMAInterleaving)) { | 3763 #endif // COMPILER1 3764 ) { 3765 if (!FLAG_IS_DEFAULT(UseBiasedLocking) && UseBiasedLocking) { 3766 // flag set to true on command line; warn the user that they 3767 // can't enable biased locking here 3768 warning("Biased Locking is not supported with locking debug flags" 3769 "; ignoring UseBiasedLocking flag." ); 3770 } 3771 UseBiasedLocking = false; 3772 } 3773 3774 #ifdef ZERO 3775 // Clear flags not supported on zero. 3776 FLAG_SET_DEFAULT(ProfileInterpreter, false); 3777 FLAG_SET_DEFAULT(UseBiasedLocking, false); 3778 LP64_ONLY(FLAG_SET_DEFAULT(UseCompressedOops, false)); 3779 LP64_ONLY(FLAG_SET_DEFAULT(UseCompressedClassPointers, false)); 3780 #endif // CC_INTERP 3781 3782 #ifdef COMPILER2 3783 if (!EliminateLocks) { 3784 EliminateNestedLocks = false; 3785 } 3786 if (!Inline) { 3787 IncrementalInline = false; 3788 } 3789 #ifndef PRODUCT 3790 if (!IncrementalInline) { 3791 AlwaysIncrementalInline = false; 3792 } 3793 #endif 3794 if (IncrementalInline && FLAG_IS_DEFAULT(MaxNodeLimit)) { 3795 // incremental inlining: bump MaxNodeLimit 3796 FLAG_SET_DEFAULT(MaxNodeLimit, (intx)75000); 3797 } 3798 if (!UseTypeSpeculation && FLAG_IS_DEFAULT(TypeProfileLevel)) { 3799 // nothing to use the profiling, turn if off 3800 FLAG_SET_DEFAULT(TypeProfileLevel, 0); 3801 } 3802 if (UseTypeSpeculation && FLAG_IS_DEFAULT(ReplaceInParentMaps)) { 3823 } 3824 3825 if (!LogVMOutput && FLAG_IS_DEFAULT(LogVMOutput)) { 3826 if (use_vm_log()) { 3827 LogVMOutput = true; 3828 } 3829 } 3830 #endif // PRODUCT 3831 3832 if (PrintCommandLineFlags) { 3833 CommandLineFlags::printSetFlags(tty); 3834 } 3835 3836 // Apply CPU specific policy for the BiasedLocking 3837 if (UseBiasedLocking) { 3838 if (!VM_Version::use_biased_locking() && 3839 !(FLAG_IS_CMDLINE(UseBiasedLocking))) { 3840 UseBiasedLocking = false; 3841 } 3842 } 3843 #ifdef COMPILER2 3844 if (!UseBiasedLocking || EmitSync != 0) { 3845 UseOptoBiasInlining = false; 3846 } 3847 #endif 3848 3849 return JNI_OK; 3850 } 3851 3852 jint Arguments::adjust_after_os() { 3853 if (UseNUMA) { 3854 if (UseParallelGC || UseParallelOldGC) { 3855 if (FLAG_IS_DEFAULT(MinHeapDeltaBytes)) { 3856 FLAG_SET_DEFAULT(MinHeapDeltaBytes, 64*M); 3857 } 3858 } 3859 // UseNUMAInterleaving is set to ON for all collectors and 3860 // platforms when UseNUMA is set to ON. NUMA-aware collectors 3861 // such as the parallel collector for Linux and Solaris will 3862 // interleave old gen and survivor spaces on top of NUMA 3863 // allocation policy for the eden space. 3864 // Non NUMA-aware collectors such as CMS, G1 and Serial-GC on 3865 // all platforms and ParallelGC on Windows will interleave all 3866 // of the heap spaces across NUMA nodes. 3867 if (FLAG_IS_DEFAULT(UseNUMAInterleaving)) { |