Mercurial > hg > graal-compiler
changeset 22299:c28cb37b2e1d
Rename JVMCI to INCLUDE_JVMCI.
line wrap: on
line diff
--- a/make/bsd/makefiles/compiler1.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/bsd/makefiles/compiler1.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER1 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = client-nojvmci endif \ No newline at end of file
--- a/make/bsd/makefiles/compiler2.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/bsd/makefiles/compiler2.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER2 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = server-nojvmci endif \ No newline at end of file
--- a/make/bsd/makefiles/tiered.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/bsd/makefiles/tiered.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER2 -DCOMPILER1 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = server-nojvmci endif
--- a/make/excludeSrc.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/excludeSrc.make Wed Jul 22 08:56:03 2015 -0700 @@ -127,6 +127,11 @@ memTracker.cpp nmtDCmd.cpp mallocSiteTable.cpp endif +ifeq ($(INCLUDE_JVMCI), false) + CXXFLAGS += -DINCLUDE_JVMCI=0 + CFLAGS += -DINCLUDE_JVMCI=0 +endif + -include $(HS_ALT_MAKE)/excludeSrc.make .PHONY: $(HS_ALT_MAKE)/excludeSrc.make
--- a/make/linux/makefiles/compiler1.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/linux/makefiles/compiler1.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER1 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = client-nojvmci endif \ No newline at end of file
--- a/make/linux/makefiles/compiler2.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/linux/makefiles/compiler2.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER2 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = server-nojvmci endif \ No newline at end of file
--- a/make/linux/makefiles/tiered.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/linux/makefiles/tiered.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER2 -DCOMPILER1 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = server-nojvmci endif \ No newline at end of file
--- a/make/solaris/makefiles/compiler1.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/compiler1.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER1 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = client-nojvmci endif
--- a/make/solaris/makefiles/compiler2.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/compiler2.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER2 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = server-nojvmci endif
--- a/make/solaris/makefiles/debug.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/debug.make Wed Jul 22 08:56:03 2015 -0700 @@ -46,9 +46,6 @@ # Linker mapfiles MAPFILE = $(GAMMADIR)/make/solaris/makefiles/mapfile-vers \ $(GAMMADIR)/make/solaris/makefiles/mapfile-vers-debug -#ifdef JVMCI -MAPFILE += $(GAMMADIR)/make/solaris/makefiles/mapfile-vers-JVMCI -#endif # This mapfile is only needed when compiling with dtrace support, # and mustn't be otherwise.
--- a/make/solaris/makefiles/fastdebug.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/fastdebug.make Wed Jul 22 08:56:03 2015 -0700 @@ -117,9 +117,6 @@ # Linker mapfiles MAPFILE = $(GAMMADIR)/make/solaris/makefiles/mapfile-vers \ $(GAMMADIR)/make/solaris/makefiles/mapfile-vers-debug -#ifdef JVMCI -MAPFILE += $(GAMMADIR)/make/solaris/makefiles/mapfile-vers-JVMCI -#endif # This mapfile is only needed when compiling with dtrace support, # and mustn't be otherwise.
--- a/make/solaris/makefiles/mapfile-vers-JVMCI Wed Jul 22 15:50:08 2015 +0200 +++ /dev/null Thu Jan 01 00:00:00 1970 +0000 @@ -1,30 +0,0 @@ -# - -# -# Copyright (c) 2006, 2008, Oracle and/or its affiliates. All rights reserved. -# DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. -# -# This code is free software; you can redistribute it and/or modify it -# under the terms of the GNU General Public License version 2 only, as -# published by the Free Software Foundation. -# -# This code is distributed in the hope that it will be useful, but WITHOUT -# ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or -# FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License -# version 2 for more details (a copy is included in the LICENSE file that -# accompanied this code). -# -# You should have received a copy of the GNU General Public License version -# 2 along with this work; if not, write to the Free Software Foundation, -# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. -# -# Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA -# or visit www.oracle.com if you need additional information or have any -# questions. -# -# - -# Define public interface. -SUNWprivate_1.1 { - global: -};
--- a/make/solaris/makefiles/optimized.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/optimized.make Wed Jul 22 08:56:03 2015 -0700 @@ -54,9 +54,6 @@ # Linker mapfiles MAPFILE = $(GAMMADIR)/make/solaris/makefiles/mapfile-vers -#ifdef JVMCI -MAPFILE += $(GAMMADIR)/make/solaris/makefiles/mapfile-vers-JVMCI -#endif # This mapfile is only needed when compiling with dtrace support, # and mustn't be otherwise.
--- a/make/solaris/makefiles/product.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/product.make Wed Jul 22 08:56:03 2015 -0700 @@ -66,9 +66,6 @@ # Linker mapfiles MAPFILE = $(GAMMADIR)/make/solaris/makefiles/mapfile-vers -ifdef JVMCI -MAPFILE += $(GAMMADIR)/make/solaris/makefiles/mapfile-vers-JVMCI -endif ifndef USE_GCC # This mapfile is only needed when compiling with dtrace support,
--- a/make/solaris/makefiles/tiered.make Wed Jul 22 15:50:08 2015 +0200 +++ b/make/solaris/makefiles/tiered.make Wed Jul 22 08:56:03 2015 -0700 @@ -30,8 +30,6 @@ CFLAGS += -DCOMPILER2 -DCOMPILER1 -ifeq ($(INCLUDE_JVMCI), true) - CFLAGS += -DJVMCI -else +ifeq ($(INCLUDE_JVMCI), false) VM_SUBDIR = server-nojvmci endif
--- a/src/cpu/sparc/vm/compiledIC_sparc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/sparc/vm/compiledIC_sparc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -102,7 +102,7 @@ void CompiledStaticCall::set_to_interpreted(methodHandle callee, address entry) { address stub = find_stub(); -#ifdef JVMCI +#if INCLUDE_JVMCI if (stub == NULL) { set_destination_mt_safe(entry); return; @@ -159,7 +159,7 @@ // Verify stub. address stub = find_stub(); -#ifndef JVMCI +#if !INCLUDE_JVMCI assert(stub != NULL, "no stub found for static call"); // Creation also verifies the object. NativeMovConstReg* method_holder = nativeMovConstReg_at(stub);
--- a/src/cpu/sparc/vm/interp_masm_sparc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/sparc/vm/interp_masm_sparc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -1644,7 +1644,7 @@ bind(skip_receiver_profile); // The method data pointer needs to be updated to reflect the new target. -#ifdef JVMCI +#if INCLUDE_JVMCI if (MethodProfileWidth == 0) { update_mdp_by_constant(in_bytes(VirtualCallData::virtual_call_data_size())); } @@ -1655,7 +1655,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI void InterpreterMacroAssembler::profile_called_method(Register method, Register scratch) { assert_different_registers(method, scratch); if (ProfileInterpreter && MethodProfileWidth > 0) { @@ -1681,7 +1681,7 @@ if (is_virtual_call) { increment_mdp_data_at(in_bytes(CounterData::count_offset()), scratch); } -#ifdef JVMCI +#if INCLUDE_JVMCI else { increment_mdp_data_at(in_bytes(ReceiverTypeData::nonprofiled_receiver_count_offset()), scratch); } @@ -1691,11 +1691,11 @@ int non_profiled_offset = -1; if (use_non_profiled_counter) { non_profiled_offset = in_bytes(CounterData::count_offset()); - #ifdef JVMCI +#if INCLUDE_JVMCI if (!is_virtual_call) { non_profiled_offset = in_bytes(ReceiverTypeData::nonprofiled_receiver_count_offset()); } - #endif +#endif assert(non_profiled_offset >= 0, "must be"); }
--- a/src/cpu/sparc/vm/sharedRuntime_sparc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/sparc/vm/sharedRuntime_sparc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -43,7 +43,7 @@ #include "compiler/compileBroker.hpp" #include "shark/sharkCompiler.hpp" #endif -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciJavaAccess.hpp" #endif @@ -995,7 +995,7 @@ // Jump to the compiled code just as if compiled code was doing it. __ ld_ptr(G5_method, in_bytes(Method::from_compiled_offset()), G3); -#ifdef JVMCI +#if INCLUDE_JVMCI // check if this call should be routed towards a specific entry point __ ld(Address(G2_thread, in_bytes(JavaThread::jvmci_alternate_call_target_offset())), G1); __ cmp(G0, G1); @@ -3473,7 +3473,7 @@ pad += StackShadowPages*16 + 32; } #endif -#ifdef JVMCI +#if INCLUDE_JVMCI pad += 1000; // Increase the buffer size when compiling for JVMCI #endif #ifdef _LP64 @@ -3543,7 +3543,7 @@ __ delayed()->mov(Deoptimization::Unpack_deopt, L0deopt_mode); -#ifdef JVMCI +#if INCLUDE_JVMCI masm->block_comment("BEGIN JVMCI"); int implicit_exception_uncommon_trap_offset = __ offset() - start; __ ld_ptr(G2_thread, in_bytes(JavaThread::jvmci_implicit_exception_pc_offset()), O7); @@ -3576,7 +3576,7 @@ __ ba(after_fetch_unroll_info_call); __ delayed()->nop(); // Delay slot masm->block_comment("END JVMCI"); -#endif // JVMCI +#endif // INCLUDE_JVMCI int exception_offset = __ offset() - start; @@ -3659,7 +3659,7 @@ __ reset_last_Java_frame(); -#ifdef JVMCI +#if INCLUDE_JVMCI __ bind(after_fetch_unroll_info_call); #endif // NOTE: we know that only O0/O1 will be reloaded by restore_result_registers @@ -3727,7 +3727,7 @@ masm->flush(); _deopt_blob = DeoptimizationBlob::create(&buffer, oop_maps, 0, exception_offset, reexecute_offset, frame_size_words); _deopt_blob->set_unpack_with_exception_in_tls_offset(exception_in_tls_offset); -#ifdef JVMCI +#if INCLUDE_JVMCI _deopt_blob->set_uncommon_trap_offset(uncommon_trap_offset); _deopt_blob->set_implicit_exception_uncommon_trap_offset(implicit_exception_uncommon_trap_offset); #endif
--- a/src/cpu/sparc/vm/templateInterpreter_sparc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/sparc/vm/templateInterpreter_sparc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -203,7 +203,7 @@ address InterpreterGenerator::generate_deopt_entry_for(TosState state, int step) { address entry = __ pc(); __ get_constant_pool_cache(LcpoolCache); // load LcpoolCache -#ifdef JVMCI +#if INCLUDE_JVMCI // Check if we need to take lock at entry of synchronized method. { Label L;
--- a/src/cpu/sparc/vm/templateTable_sparc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/sparc/vm/templateTable_sparc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -2949,7 +2949,7 @@ // get target Method* & entry point __ lookup_virtual_method(Rrecv, Rindex, G5_method); __ profile_arguments_type(G5_method, Rcall, Gargs, true); -#ifdef JVMCI +#if INCLUDE_JVMCI __ profile_called_method(G5_method, Rtemp); #endif __ call_from_interpreter(Rcall, Gargs, Rret); @@ -3206,7 +3206,7 @@ assert_different_registers(Rcall, G5_method, Gargs, Rret); __ profile_arguments_type(G5_method, Rcall, Gargs, true); -#ifdef JVMCI +#if INCLUDE_JVMCI __ profile_called_method(G5_method, Rscratch); #endif __ call_from_interpreter(Rcall, Gargs, Rret);
--- a/src/cpu/x86/vm/c2_globals_x86.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/c2_globals_x86.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -86,7 +86,7 @@ define_pd_global(bool, OptoScheduling, false); define_pd_global(bool, OptoBundling, false); -#ifdef JVMCI +#if INCLUDE_JVMCI define_pd_global(intx, ReservedCodeCacheSize, 64*M); #else define_pd_global(intx, ReservedCodeCacheSize, 48*M);
--- a/src/cpu/x86/vm/frame_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/frame_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -460,11 +460,11 @@ // This is the sp before any possible extension (adapter/locals). intptr_t* unextended_sp = interpreter_frame_sender_sp(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (map->update_map()) { update_map_with_saved_link(map, (intptr_t**) addr_at(link_offset)); } -#endif // COMPILER2 || JVMCI +#endif // COMPILER2 || INCLUDE_JVMCI return frame(sender_sp, unextended_sp, link(), sender_pc()); }
--- a/src/cpu/x86/vm/globals_x86.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/globals_x86.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -47,7 +47,7 @@ // the the vep is aligned at CodeEntryAlignment whereas c2 only aligns // the uep and the vep doesn't get real alignment but just slops on by // only assured that the entry instruction meets the 5 byte size requirement. -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI define_pd_global(intx, CodeEntryAlignment, 32); #else define_pd_global(intx, CodeEntryAlignment, 16);
--- a/src/cpu/x86/vm/interp_masm_x86_64.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/interp_masm_x86_64.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -1129,7 +1129,7 @@ bind(skip_receiver_profile); // The method data pointer needs to be updated to reflect the new target. -#ifdef JVMCI +#if INCLUDE_JVMCI if (MethodProfileWidth == 0) { update_mdp_by_constant(mdp, in_bytes(VirtualCallData::virtual_call_data_size())); } @@ -1140,7 +1140,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI void InterpreterMacroAssembler::profile_called_method(Register method, Register mdp, Register reg2) { assert_different_registers(method, mdp, reg2); if (ProfileInterpreter && MethodProfileWidth > 0) { @@ -1178,7 +1178,7 @@ if (is_virtual_call) { increment_mdp_data_at(mdp, in_bytes(CounterData::count_offset())); } -#ifdef JVMCI +#if INCLUDE_JVMCI else { increment_mdp_data_at(mdp, in_bytes(ReceiverTypeData::nonprofiled_receiver_count_offset())); } @@ -1188,11 +1188,11 @@ int non_profiled_offset = -1; if (use_non_profiled_counter) { non_profiled_offset = in_bytes(CounterData::count_offset()); - #ifdef JVMCI +#if INCLUDE_JVMCI if (!is_virtual_call) { non_profiled_offset = in_bytes(ReceiverTypeData::nonprofiled_receiver_count_offset()); } - #endif +#endif assert(non_profiled_offset >= 0, "must be"); }
--- a/src/cpu/x86/vm/macroAssembler_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/macroAssembler_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -2945,7 +2945,7 @@ } // !defined(COMPILER2) is because of stupid core builds -#if !defined(_LP64) || defined(COMPILER1) || !defined(COMPILER2) || defined(JVMCI) +#if !defined(_LP64) || defined(COMPILER1) || !defined(COMPILER2) || INCLUDE_JVMCI void MacroAssembler::empty_FPU_stack() { if (VM_Version::supports_mmx()) { emms(); @@ -2953,7 +2953,7 @@ for (int i = 8; i-- > 0; ) ffree(i); } } -#endif // !LP64 || C1 || !C2 +#endif // !LP64 || C1 || !C2 || JVMCI // Defines obj, preserves var_size_in_bytes
--- a/src/cpu/x86/vm/relocInfo_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/relocInfo_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -181,7 +181,7 @@ #ifdef _LP64 typedef Assembler::WhichOperand WhichOperand; WhichOperand which = (WhichOperand) format(); -#ifndef JVMCI +#if !INCLUDE_JVMCI assert((which == Assembler::disp32_operand) == !Assembler::is_polling_page_far(), "format not set correctly"); #endif if (which == Assembler::disp32_operand) {
--- a/src/cpu/x86/vm/sharedRuntime_x86_64.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/sharedRuntime_x86_64.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -40,7 +40,7 @@ #ifdef COMPILER2 #include "opto/runtime.hpp" #endif -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciJavaAccess.hpp" #endif @@ -141,7 +141,7 @@ OopMap* RegisterSaver::save_live_registers(MacroAssembler* masm, int additional_frame_words, int* total_frame_words, bool save_vectors) { int vect_words = 0; int ymmhi_offset = -1; -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (save_vectors) { assert(UseAVX > 0, "256bit vectors are supported only with AVX"); assert(MaxVectorSize == 32, "only 256bit vectors are supported now"); @@ -244,7 +244,7 @@ map->set_callee_saved(STACK_OFFSET(xmm15_off), xmm15->as_VMReg()); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (save_vectors) { assert(ymmhi_offset != -1, "save area must exist"); map->set_callee_saved(YMMHI_STACK_OFFSET( 0), xmm0->as_VMReg()->next()->next()->next()->next()); @@ -309,7 +309,7 @@ // Pop arg register save area __ addptr(rsp, frame::arg_reg_save_area_bytes); } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (restore_vectors) { // Restore upper half of YMM registes. assert(UseAVX > 0, "256bit vectors are supported only with AVX"); @@ -731,7 +731,7 @@ __ block_comment("} verify_i2ce "); } -#ifdef JVMCI +#if INCLUDE_JVMCI if (frame_extension_argument != -1) { // The frame_extension_argument is an int that describes the // expected amount of argument space in the caller frame. If that @@ -802,7 +802,7 @@ // Pre-load the register-jump target early, to schedule it better. __ movptr(r11, Address(rbx, in_bytes(Method::from_compiled_offset()))); -#ifdef JVMCI +#if INCLUDE_JVMCI // check if this call should be routed towards a specific entry point __ cmpptr(Address(r15_thread, in_bytes(JavaThread::jvmci_alternate_call_target_offset())), 0); Label no_alternative_target; @@ -3430,7 +3430,7 @@ __ movl(r14, Deoptimization::Unpack_reexecute); // callee-saved __ jmp(cont); -#ifdef JVMCI +#if INCLUDE_JVMCI int implicit_exception_uncommon_trap_offset = __ pc() - start; __ pushptr(Address(r15_thread, in_bytes(JavaThread::jvmci_implicit_exception_pc_offset()))); @@ -3454,7 +3454,7 @@ Label after_fetch_unroll_info_call; __ jmp(after_fetch_unroll_info_call); -#endif // JVMCI +#endif // INCLUDE_JVMCI int exception_offset = __ pc() - start; @@ -3541,7 +3541,7 @@ __ reset_last_Java_frame(false, false); -#ifdef JVMCI +#if INCLUDE_JVMCI __ bind(after_fetch_unroll_info_call); #endif @@ -3719,7 +3719,7 @@ _deopt_blob = DeoptimizationBlob::create(&buffer, oop_maps, 0, exception_offset, reexecute_offset, frame_size_in_words); _deopt_blob->set_unpack_with_exception_in_tls_offset(exception_in_tls_offset); -#ifdef JVMCI +#if INCLUDE_JVMCI _deopt_blob->set_uncommon_trap_offset(uncommon_trap_offset); _deopt_blob->set_implicit_exception_uncommon_trap_offset(implicit_exception_uncommon_trap_offset); #endif
--- a/src/cpu/x86/vm/templateInterpreter_x86_64.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/templateInterpreter_x86_64.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -204,7 +204,7 @@ __ movptr(Address(rbp, frame::interpreter_frame_last_sp_offset * wordSize), (int32_t)NULL_WORD); __ restore_bcp(); __ restore_locals(); -#ifdef JVMCI +#if INCLUDE_JVMCI // Check if we need to take lock at entry of synchronized method. { Label L;
--- a/src/cpu/x86/vm/templateTable_x86_64.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/templateTable_x86_64.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -3039,7 +3039,7 @@ // get target Method* & entry point __ lookup_virtual_method(rax, index, method); -#ifdef JVMCI +#if INCLUDE_JVMCI // r14: MethodDataPointer (r14 is callee saved) __ profile_called_method(method, r14, r13); #endif @@ -3143,7 +3143,7 @@ __ testptr(rbx, rbx); __ jcc(Assembler::zero, no_such_method); -#ifdef JVMCI +#if INCLUDE_JVMCI // r13: MethodDataPointer (r13 is callee saved) __ profile_called_method(rbx, r13, r14); #endif
--- a/src/cpu/x86/vm/vm_version_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/cpu/x86/vm/vm_version_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -668,7 +668,7 @@ } } #endif -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (MaxVectorSize > 0) { if (!is_power_of_2(MaxVectorSize)) { warning("MaxVectorSize must be a power of 2");
--- a/src/os/windows/vm/os_windows.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/os/windows/vm/os_windows.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -2253,7 +2253,7 @@ #elif _M_AMD64 PCONTEXT ctx = exceptionInfo->ContextRecord; address pc = (address)ctx->Rip; -#ifdef JVMCI +#if INCLUDE_JVMCI assert(pc[0] >= Assembler::REX && pc[0] <= Assembler::REX_WRXB && pc[1] == 0xF7 || pc[0] == 0xF7, "not an idiv opcode"); if (pc[0] == 0xF7) { // set correct result values and continue after idiv instruction
--- a/src/os_cpu/bsd_x86/vm/thread_bsd_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/os_cpu/bsd_x86/vm/thread_bsd_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -66,7 +66,7 @@ frame ret_frame(ret_sp, ret_fp, addr.pc()); if (!ret_frame.safe_for_sender(jt)) { -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // C2 and JVMCI use ebp as a general register see if NULL fp helps frame ret_frame2(ret_sp, NULL, addr.pc()); if (!ret_frame2.safe_for_sender(jt)) { @@ -77,7 +77,7 @@ #else // nothing else to try if the frame isn't good return false; -#endif /* COMPILER2 || JVMCI*/ +#endif /* COMPILER2 || INCLUDE_JVMCI */ } *fr_addr = ret_frame; return true;
--- a/src/os_cpu/linux_x86/vm/thread_linux_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/os_cpu/linux_x86/vm/thread_linux_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -67,7 +67,7 @@ frame ret_frame(ret_sp, ret_fp, addr.pc()); if (!ret_frame.safe_for_sender(jt)) { -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // C2 and JVMCI use ebp as a general register see if NULL fp helps frame ret_frame2(ret_sp, NULL, addr.pc()); if (!ret_frame2.safe_for_sender(jt)) { @@ -78,7 +78,7 @@ #else // nothing else to try if the frame isn't good return false; -#endif /* COMPILER2 || JVMCI */ +#endif /* COMPILER2 || INCLUDE_JVMCI */ } *fr_addr = ret_frame; return true;
--- a/src/os_cpu/solaris_sparc/vm/os_solaris_sparc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/os_cpu/solaris_sparc/vm/os_solaris_sparc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -484,14 +484,14 @@ } #endif // COMPILER2 -#ifdef JVMCI +#if INCLUDE_JVMCI else if (sig == SIGILL && info->si_trapno == 0x100 + ST_RESERVED_FOR_USER_0) { - printf("SIGILL 0x%x 0x%x\n", info->si_trapno, ST_RESERVED_FOR_USER_0); - uc->uc_mcontext.gregs[REG_PC] = (greg_t)pc+4; - uc->uc_mcontext.gregs[REG_nPC] = (greg_t)npc + 4; - return true; + printf("SIGILL 0x%x 0x%x\n", info->si_trapno, ST_RESERVED_FOR_USER_0); + uc->uc_mcontext.gregs[REG_PC] = (greg_t)pc+4; + uc->uc_mcontext.gregs[REG_nPC] = (greg_t)npc + 4; + return true; } -#endif // JVMCI +#endif else if (sig == SIGSEGV && info->si_code > 0 && !MacroAssembler::needs_explicit_null_check((intptr_t)info->si_addr)) { // Determination of interpreter/vtable stub/compiled code null exception
--- a/src/os_cpu/windows_x86/vm/thread_windows_x86.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/os_cpu/windows_x86/vm/thread_windows_x86.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -74,7 +74,7 @@ frame ret_frame(ret_sp, ret_fp, addr.pc()); if (!ret_frame.safe_for_sender(jt)) { -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // C2 and JVMCI use ebp as a general register see if NULL fp helps frame ret_frame2(ret_sp, NULL, addr.pc()); if (!ret_frame2.safe_for_sender(jt)) { @@ -85,7 +85,7 @@ #else // nothing else to try if the frame isn't good return false; -#endif /* COMPILER2 || JVMCI */ +#endif /* COMPILER2 || INCLUDE_JVMCI */ } *fr_addr = ret_frame; return true;
--- a/src/share/vm/classfile/classFileParser.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/classfile/classFileParser.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -1791,7 +1791,7 @@ if (_location != _in_method) break; // only allow for methods if (!privileged) break; // only allow in privileged code return _method_LambdaForm_Hidden; -#ifdef JVMCI +#if INCLUDE_JVMCI case vmSymbols::VM_SYMBOL_ENUM_NAME(jdk_internal_jvmci_hotspot_Stable_signature): if (_location != _in_field) break; // only allow for fields if (!privileged && loader_data->class_loader() != SystemDictionary::jvmci_loader()) break; // only allow in privileged code
--- a/src/share/vm/classfile/javaClasses.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/classfile/javaClasses.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -563,7 +563,7 @@ set_init_lock(mirror(), r); // Set protection domain also -#ifdef JVMCI +#if INCLUDE_JVMCI if (k->class_loader() == SystemDictionary::jvmci_loader()) { // Same protection domain as for classes loaded by the boot loader protection_domain = Handle();
--- a/src/share/vm/classfile/systemDictionary.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/classfile/systemDictionary.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -64,7 +64,7 @@ #include "services/threadService.hpp" #include "utilities/macros.hpp" #include "utilities/ticks.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciRuntime.hpp" #endif #if INCLUDE_TRACE @@ -96,7 +96,7 @@ bool SystemDictionary::_has_loadClassInternal = false; bool SystemDictionary::_has_checkPackageAccess = false; -#ifdef JVMCI +#if INCLUDE_JVMCI oop SystemDictionary::_jvmci_loader = NULL; oop SystemDictionary::jvmci_loader() { @@ -1865,7 +1865,7 @@ Klass** klassp = &_well_known_klasses[id]; bool must_load = (init_opt < SystemDictionary::Opt); if ((*klassp) == NULL) { -#ifdef JVMCI +#if INCLUDE_JVMCI bool is_jvmci = init_opt == SystemDictionary::Jvmci; assert(is_jvmci == (id >= (int)FIRST_JVMCI_WKID && id <= (int)LAST_JVMCI_WKID), "JVMCI WKIDs must be contiguous and separate from non-JVMCI WKIDs");
--- a/src/share/vm/classfile/systemDictionary.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/classfile/systemDictionary.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -268,7 +268,7 @@ WKID_LIMIT, -#ifdef JVMCI +#if INCLUDE_JVMCI FIRST_JVMCI_WKID = WK_KLASS_ENUM_NAME(Debug_klass), LAST_JVMCI_WKID = WK_KLASS_ENUM_NAME(AbstractValue_klass), #endif @@ -286,7 +286,7 @@ Opt, // preload tried; NULL if not present Opt_Only_JDK14NewRef, // preload tried; use only with NewReflection Opt_Only_JDK15, // preload tried; use only with JDK1.5+ -#ifdef JVMCI +#if INCLUDE_JVMCI Jvmci, // preload tried; error if not present, use only with JVMCI #endif OPTION_LIMIT, @@ -478,7 +478,7 @@ // despite the optional loading, if you use this it must be present: return check_klass(k); } -#ifdef JVMCI +#if INCLUDE_JVMCI static Klass* check_klass_Jvmci(Klass* k) { return k; } #endif @@ -546,7 +546,7 @@ // Returns default system loader static oop java_system_loader(); -#ifdef JVMCI +#if INCLUDE_JVMCI // Returns the JVMCI loader. This will be NULL if !UseJVMCIClassLoader // in which case it's equivalent to the boot loader static oop jvmci_loader(); @@ -784,7 +784,7 @@ static Klass* _box_klasses[T_VOID+1]; static oop _java_system_loader; -#ifdef JVMCI +#if INCLUDE_JVMCI static oop _jvmci_loader; #endif
--- a/src/share/vm/code/codeBlob.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/codeBlob.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -357,7 +357,7 @@ int _unpack_with_exception_in_tls; -#ifdef JVMCI +#if INCLUDE_JVMCI // (thomaswue) Offset when JVMCI calls uncommon_trap. int _uncommon_trap_offset; int _implicit_exception_uncommon_trap_offset; @@ -416,7 +416,7 @@ } address unpack_with_exception_in_tls() const { return code_begin() + _unpack_with_exception_in_tls; } -#ifdef JVMCI +#if INCLUDE_JVMCI // (thomaswue) Offset when JVMCI calls uncommon_trap. void set_uncommon_trap_offset(int offset) { _uncommon_trap_offset = offset;
--- a/src/share/vm/code/debugInfoRec.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/debugInfoRec.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -37,7 +37,7 @@ int _offset; // location in the stream of this scope int _length; // number of bytes in the stream int _hash; // hash of stream bytes (for quicker reuse) -#ifdef JVMCI +#if INCLUDE_JVMCI DebugInformationRecorder* _DIR; #endif @@ -54,7 +54,7 @@ DIR_Chunk(int offset, int length, DebugInformationRecorder* dir) { _offset = offset; _length = length; -#ifdef JVMCI +#if INCLUDE_JVMCI _DIR = dir; #endif unsigned int hash = 0; @@ -84,7 +84,7 @@ return NULL; } -#ifdef JVMCI +#if INCLUDE_JVMCI static int compare(DIR_Chunk* a, DIR_Chunk* b) { if (b->_hash > a->_hash) { return 1; @@ -138,7 +138,7 @@ _oop_recorder = oop_recorder; _all_chunks = new GrowableArray<DIR_Chunk*>(300); -#ifndef JVMCI +#if !INCLUDE_JVMCI _shared_chunks = new GrowableArray<DIR_Chunk*>(30); #endif _next_chunk = _next_chunk_limit = NULL; @@ -277,7 +277,7 @@ DIR_Chunk* ns = new(this) DIR_Chunk(stream_offset, stream_length, this); -#ifdef JVMCI +#if INCLUDE_JVMCI DIR_Chunk* match = _all_chunks->find_insert_binary<DIR_Chunk::compare>(ns); if (match != ns) { // Found an existing chunk
--- a/src/share/vm/code/debugInfoRec.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/debugInfoRec.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -163,7 +163,7 @@ // Scopes that have been described so far. GrowableArray<DIR_Chunk*>* _all_chunks; -#ifndef JVMCI +#if !INCLUDE_JVMCI GrowableArray<DIR_Chunk*>* _shared_chunks; #endif DIR_Chunk* _next_chunk;
--- a/src/share/vm/code/dependencies.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/dependencies.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -52,9 +52,9 @@ _oop_recorder = env->oop_recorder(); _log = env->log(); _dep_seen = new(arena) GrowableArray<int>(arena, 500, 0, 0); -#ifdef JVMCI +#if INCLUDE_JVMCI _using_dep_values = false; -#endif // JVMCI +#endif DEBUG_ONLY(_deps[end_marker] = NULL); for (int i = (int)FIRST_TYPE; i < (int)TYPE_LIMIT; i++) { _deps[i] = new(arena) GrowableArray<ciBaseObject*>(arena, 10, 0, 0); @@ -124,7 +124,7 @@ assert_common_2(call_site_target_value, call_site, method_handle); } -#ifdef JVMCI +#if INCLUDE_JVMCI Dependencies::Dependencies(Arena* arena, OopRecorder* oop_recorder, CompileLog* log) { _oop_recorder = oop_recorder; @@ -180,7 +180,7 @@ void Dependencies::assert_call_site_target_value(oop call_site, oop method_handle) { assert_common_2(call_site_target_value, DepValue(_oop_recorder, JNIHandles::make_local(call_site)), DepValue(_oop_recorder, JNIHandles::make_local(method_handle))); } -#endif // JVMCI +#endif // INCLUDE_JVMCI // Helper function. If we are adding a new dep. under ctxk2, @@ -294,7 +294,7 @@ deps->append(x2); } -#ifdef JVMCI +#if INCLUDE_JVMCI bool Dependencies::maybe_merge_ctxk(GrowableArray<DepValue>* deps, int ctxk_i, DepValue ctxk2_dv) { Klass* ctxk1 = deps->at(ctxk_i).as_klass(_oop_recorder); @@ -364,7 +364,7 @@ deps->append(x0); deps->append(x1); } -#endif // JVMCI +#endif // INCLUDE_JVMCI /// Support for encoding dependencies into an nmethod: @@ -392,7 +392,7 @@ static int sort_dep_arg_3(ciBaseObject** p1, ciBaseObject** p2) { return sort_dep(p1, p2, 3); } -#ifdef JVMCI +#if INCLUDE_JVMCI // metadata deps are sorted before object deps static int sort_dep_value(Dependencies::DepValue* p1, Dependencies::DepValue* p2, int narg) { for (int i = 0; i < narg; i++) { @@ -407,10 +407,10 @@ { return sort_dep_value(p1, p2, 2); } static int sort_dep_value_arg_3(Dependencies::DepValue* p1, Dependencies::DepValue* p2) { return sort_dep_value(p1, p2, 3); } -#endif // JVMCI +#endif // INCLUDE_JVMCI void Dependencies::sort_all_deps() { -#ifdef JVMCI +#if INCLUDE_JVMCI if (_using_dep_values) { for (int deptv = (int)FIRST_TYPE; deptv < (int)TYPE_LIMIT; deptv++) { DepType dept = (DepType)deptv; @@ -425,7 +425,7 @@ } return; } -#endif // JVMCI +#endif for (int deptv = (int)FIRST_TYPE; deptv < (int)TYPE_LIMIT; deptv++) { DepType dept = (DepType)deptv; GrowableArray<ciBaseObject*>* deps = _deps[dept]; @@ -441,7 +441,7 @@ size_t Dependencies::estimate_size_in_bytes() { size_t est_size = 100; -#ifdef JVMCI +#if INCLUDE_JVMCI if (_using_dep_values) { for (int deptv = (int)FIRST_TYPE; deptv < (int)TYPE_LIMIT; deptv++) { DepType dept = (DepType)deptv; @@ -450,7 +450,7 @@ } return est_size; } -#endif // JVMCI +#endif for (int deptv = (int)FIRST_TYPE; deptv < (int)TYPE_LIMIT; deptv++) { DepType dept = (DepType)deptv; GrowableArray<ciBaseObject*>* deps = _deps[dept]; @@ -490,7 +490,7 @@ // cast is safe, no deps can overflow INT_MAX CompressedWriteStream bytes((int)estimate_size_in_bytes()); -#ifdef JVMCI +#if INCLUDE_JVMCI if (_using_dep_values) { for (int deptv = (int)FIRST_TYPE; deptv < (int)TYPE_LIMIT; deptv++) { DepType dept = (DepType)deptv; @@ -520,7 +520,7 @@ } } } else { -#endif // JVMCI +#endif for (int deptv = (int)FIRST_TYPE; deptv < (int)TYPE_LIMIT; deptv++) { DepType dept = (DepType)deptv; GrowableArray<ciBaseObject*>* deps = _deps[dept]; @@ -554,9 +554,9 @@ } } } -#ifdef JVMCI +#if INCLUDE_JVMCI } -#endif // JVMCI +#endif // write a sentinel byte to mark the end bytes.write_byte(end_marker);
--- a/src/share/vm/code/dependencies.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/dependencies.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -200,7 +200,7 @@ static void check_valid_dependency_type(DepType dept); -#ifdef JVMCI +#if INCLUDE_JVMCI // A Metadata* or object value recorded in an OopRecorder class DepValue VALUE_OBJ_CLASS_SPEC { private: @@ -249,7 +249,7 @@ // State for writing a new set of dependencies: GrowableArray<int>* _dep_seen; // (seen[h->ident] & (1<<dept)) GrowableArray<ciBaseObject*>* _deps[TYPE_LIMIT]; -#ifdef JVMCI +#if INCLUDE_JVMCI bool _using_dep_values; GrowableArray<DepValue>* _dep_values[TYPE_LIMIT]; #endif @@ -271,7 +271,7 @@ return (seen & (1<<dept)) != 0; } -#ifdef JVMCI +#if INCLUDE_JVMCI bool note_dep_seen(int dept, DepValue x) { assert(dept < BitsPerInt, "oops"); // place metadata deps at even indexes, object deps at odd indexes @@ -286,7 +286,7 @@ bool maybe_merge_ctxk(GrowableArray<ciBaseObject*>* deps, int ctxk_i, ciKlass* ctxk); -#ifdef JVMCI +#if INCLUDE_JVMCI bool maybe_merge_ctxk(GrowableArray<DepValue>* deps, int ctxk_i, DepValue ctxk); #endif @@ -311,9 +311,9 @@ Dependencies(ciEnv* env) { initialize(env); } -#ifdef JVMCI +#if INCLUDE_JVMCI Dependencies(Arena* arena, OopRecorder* oop_recorder, CompileLog* log); -#endif // JVMCI +#endif private: // Check for a valid context type. @@ -346,7 +346,7 @@ void assert_has_no_finalizable_subclasses(ciKlass* ctxk); void assert_call_site_target_value(ciCallSite* call_site, ciMethodHandle* method_handle); -#ifdef JVMCI +#if INCLUDE_JVMCI private: static void check_ctxk(Klass* ctxk) { assert(ctxk->oop_is_instance(), "java types only");
--- a/src/share/vm/code/nmethod.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/nmethod.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -61,7 +61,7 @@ #ifdef SHARK #include "shark/sharkCompiler.hpp" #endif -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciJavaAccess.hpp" #endif @@ -253,7 +253,7 @@ #ifdef COMPILER2 static java_nmethod_stats_struct c2_java_nmethod_stats; #endif -#ifdef JVMCI +#if INCLUDE_JVMCI static java_nmethod_stats_struct jvmci_java_nmethod_stats; #endif #ifdef SHARK @@ -279,7 +279,7 @@ c2_java_nmethod_stats.note_nmethod(nm); } else #endif -#ifdef JVMCI +#if INCLUDE_JVMCI if (nm->is_compiled_by_jvmci()) { jvmci_java_nmethod_stats.note_nmethod(nm); } else @@ -574,7 +574,7 @@ #if INCLUDE_RTM_OPT _rtm_state = NoRTM; #endif -#ifdef JVMCI +#if INCLUDE_JVMCI _jvmci_installed_code = NULL; _speculation_log = NULL; #endif @@ -672,7 +672,7 @@ ImplicitExceptionTable* nul_chk_table, AbstractCompiler* compiler, int comp_level -#ifdef JVMCI +#if INCLUDE_JVMCI , Handle installed_code, Handle speculationLog #endif @@ -699,7 +699,7 @@ nul_chk_table, compiler, comp_level -#ifdef JVMCI +#if INCLUDE_JVMCI , installed_code, speculationLog #endif @@ -935,7 +935,7 @@ ImplicitExceptionTable* nul_chk_table, AbstractCompiler* compiler, int comp_level -#ifdef JVMCI +#if INCLUDE_JVMCI , Handle installed_code, Handle speculation_log #endif @@ -963,7 +963,7 @@ _consts_offset = content_offset() + code_buffer->total_offset_of(code_buffer->consts()); _stub_offset = content_offset() + code_buffer->total_offset_of(code_buffer->stubs()); -#ifdef JVMCI +#if INCLUDE_JVMCI _jvmci_installed_code = installed_code(); _speculation_log = (instanceOop)speculation_log(); @@ -996,7 +996,7 @@ _deoptimize_mh_offset = _stub_offset + offsets->value(CodeOffsets::DeoptMH); } else { _deoptimize_mh_offset = -1; -#ifdef JVMCI +#if INCLUDE_JVMCI } #endif } @@ -1490,7 +1490,7 @@ // Unregister must be done before the state change Universe::heap()->unregister_nmethod(this); -#ifdef JVMCI +#if INCLUDE_JVMCI // The method can only be unloaded after the pointer to the installed code // Java wrapper is no longer alive. Here we need to clear out this weak // reference to the dead object. Nulling out the reference has to happen @@ -1667,7 +1667,7 @@ } else { assert(state == not_entrant, "other cases may need to be handled differently"); } -#ifdef JVMCI +#if INCLUDE_JVMCI if (_jvmci_installed_code != NULL) { // Break the link between nmethod and InstalledCode such that the nmethod can subsequently be flushed safely. InstalledCode::set_address(_jvmci_installed_code, 0); @@ -1979,7 +1979,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI // Follow JVMCI method BarrierSet* bs = Universe::heap()->barrier_set(); if (_jvmci_installed_code != NULL) { @@ -2116,7 +2116,7 @@ unloading_occurred = true; } -#ifdef JVMCI +#if INCLUDE_JVMCI // Follow JVMCI method if (_jvmci_installed_code != NULL) { if (_jvmci_installed_code->is_a(HotSpotNmethod::klass()) && HotSpotNmethod::isDefault(_jvmci_installed_code)) { @@ -2205,7 +2205,7 @@ return postponed; } -#ifdef JVMCI +#if INCLUDE_JVMCI // Follow JVMCI method BarrierSet* bs = Universe::heap()->barrier_set(); if (_jvmci_installed_code != NULL) { @@ -2415,7 +2415,7 @@ // (See comment above.) } -#ifdef JVMCI +#if INCLUDE_JVMCI if (_jvmci_installed_code != NULL) { f->do_oop((oop*) &_jvmci_installed_code); } @@ -2628,9 +2628,9 @@ // When using JVMCI the address might be off by the size of a call instruction. bool nmethod::is_deopt_entry(address pc) { return pc == deopt_handler_begin() -#ifdef JVMCI +#if INCLUDE_JVMCI || pc == (deopt_handler_begin() + NativeCall::instruction_size) -#endif // JVMCI +#endif ; } @@ -3533,7 +3533,7 @@ #ifdef COMPILER2 c2_java_nmethod_stats.print_nmethod_stats("C2"); #endif -#ifdef JVMCI +#if INCLUDE_JVMCI jvmci_java_nmethod_stats.print_nmethod_stats("JVMCI"); #endif #ifdef SHARK @@ -3548,7 +3548,7 @@ if (xtty != NULL) xtty->tail("statistics"); } -#ifdef JVMCI +#if INCLUDE_JVMCI char* nmethod::jvmci_installed_code_name(char* buf, size_t buflen) { if (!this->is_compiled_by_jvmci()) { return NULL;
--- a/src/share/vm/code/nmethod.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/nmethod.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -126,7 +126,7 @@ int _entry_bci; // != InvocationEntryBci if this nmethod is an on-stack replacement method jmethodID _jmethod_id; // Cache of method()->jmethod_id() -#ifdef JVMCI +#if INCLUDE_JVMCI // Needed to keep nmethods alive that are not the default nmethod for the associated Method. oop _jvmci_installed_code; oop _speculation_log; @@ -299,7 +299,7 @@ ImplicitExceptionTable* nul_chk_table, AbstractCompiler* compiler, int comp_level -#ifdef JVMCI +#if INCLUDE_JVMCI , Handle installed_code, Handle speculation_log #endif @@ -340,7 +340,7 @@ ImplicitExceptionTable* nul_chk_table, AbstractCompiler* compiler, int comp_level -#ifdef JVMCI +#if INCLUDE_JVMCI , Handle installed_code = Handle(), Handle speculation_log = Handle() #endif @@ -626,7 +626,7 @@ // Evolution support. We make old (discarded) compiled methods point to new Method*s. void set_method(Method* method) { _method = method; } -#ifdef JVMCI +#if INCLUDE_JVMCI oop jvmci_installed_code() { return _jvmci_installed_code ; } char* jvmci_installed_code_name(char* buf, size_t buflen); void set_jvmci_installed_code(oop installed_code) { _jvmci_installed_code = installed_code; }
--- a/src/share/vm/code/scopeDesc.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/code/scopeDesc.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -226,7 +226,7 @@ } } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (NOT_JVMCI(DoEscapeAnalysis &&) is_top() && _objects != NULL) { tty->print_cr("Objects"); for (int i = 0; i < _objects->length(); i++) { @@ -237,7 +237,7 @@ tty->cr(); } } -#endif // COMPILER2 || JVMCI +#endif // COMPILER2 || INCLUDE_JVMCI } #endif
--- a/src/share/vm/compiler/abstractCompiler.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/compiler/abstractCompiler.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -29,7 +29,7 @@ typedef void (*initializer)(void); -#ifdef JVMCI +#if INCLUDE_JVMCI // Per-compiler statistics class CompilerStatistics VALUE_OBJ_CLASS_SPEC { friend class VMStructs; @@ -89,7 +89,7 @@ private: Type _type; -#ifdef JVMCI +#if INCLUDE_JVMCI CompilerStatistics _stats; #endif @@ -136,7 +136,7 @@ ShouldNotReachHere(); } -#ifdef JVMCI +#if INCLUDE_JVMCI CompilerStatistics* stats() { return &_stats; } #endif };
--- a/src/share/vm/compiler/compileBroker.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/compiler/compileBroker.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -49,7 +49,7 @@ #ifdef COMPILER1 #include "c1/c1_Compiler.hpp" #endif -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciCompiler.hpp" #ifdef COMPILERJVMCI #include "jvmci/jvmciRuntime.hpp" @@ -916,7 +916,7 @@ // Set the interface to the current compiler(s). int c1_count = CompilationPolicy::policy()->compiler_count(CompLevel_simple); int c2_count = CompilationPolicy::policy()->compiler_count(CompLevel_full_optimization); -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCICompiler* jvmci = new JVMCICompiler(); #endif @@ -2473,7 +2473,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI void CompileBroker::print_times(AbstractCompiler* comp) { CompilerStatistics* stats = comp->stats(); tty->print_cr(" %s {speed: %d bytes/s; standard: %6.3f s, %d bytes, %d methods; osr: %6.3f s, %d bytes, %d methods; nmethods_size: %d bytes; nmethods_code_size: %d bytes}", @@ -2486,7 +2486,7 @@ #endif void CompileBroker::print_times(bool per_compiler, bool aggregate) { -#ifdef JVMCI +#if INCLUDE_JVMCI elapsedTimer standard_compilation; elapsedTimer total_compilation; elapsedTimer osr_compilation;
--- a/src/share/vm/compiler/compileBroker.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/compiler/compileBroker.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -456,7 +456,7 @@ // Redefine Classes support static void mark_on_stack(); -#ifdef JVMCI +#if INCLUDE_JVMCI // Print curent compilation time stats for a given compiler static void print_times(AbstractCompiler* comp); #endif
--- a/src/share/vm/compiler/disassembler.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/compiler/disassembler.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -529,7 +529,7 @@ nm->method()->method_holder()->name()->print_symbol_on(env.output()); env.output()->print("."); nm->method()->name()->print_symbol_on(env.output()); -#ifdef JVMCI +#if INCLUDE_JVMCI { char buffer[O_BUFLEN]; char* jvmciName = nm->jvmci_installed_code_name(buffer, O_BUFLEN);
--- a/src/share/vm/compiler/oopMap.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/compiler/oopMap.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -319,7 +319,7 @@ COMPILER1_PRESENT(ShouldNotReachHere();) COMPILERJVMCI_PRESENT(ShouldNotReachHere();) #endif // TIERED -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::add(derived, base); #endif // COMPILER2 || JVMCI } @@ -509,7 +509,7 @@ COMPILER1_PRESENT(return false); COMPILERJVMCI_PRESENT(return false); #endif // !TIERED -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI OopMapStream oms((OopMap*)this,OopMapValue::derived_oop_value); return oms.is_done(); #else @@ -581,7 +581,7 @@ //------------------------------DerivedPointerTable--------------------------- -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI class DerivedPointerEntry : public CHeapObj<mtCompiler> { private:
--- a/src/share/vm/compiler/oopMap.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/compiler/oopMap.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -283,7 +283,7 @@ // oops, it is filled in with references to all locations that contains a // derived oop (assumed to be very few). When the GC is complete, the derived // pointers are updated based on their base pointers new value and an offset. -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI class DerivedPointerTable : public AllStatic { friend class VMStructs; private:
--- a/src/share/vm/gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_implementation/concurrentMarkSweep/concurrentMarkSweepGeneration.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -2547,7 +2547,7 @@ if (UseAdaptiveSizePolicy) { size_policy()->ms_collection_begin(); } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTableDeactivate dpt_deact; #endif @@ -3023,7 +3023,7 @@ // way with the marking information used by GC. NoRefDiscovery no_discovery(ref_processor()); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTableDeactivate dpt_deact; #endif @@ -3733,7 +3733,7 @@ } { -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTableDeactivate dpt_deact; #endif if (CMSParallelInitialMarkEnabled && CollectedHeap::use_parallel_gc_threads()) { @@ -5125,7 +5125,7 @@ } { -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTableDeactivate dpt_deact; #endif
--- a/src/share/vm/gc_implementation/g1/g1CollectedHeap.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_implementation/g1/g1CollectedHeap.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -1333,7 +1333,7 @@ check_bitmaps("Full GC Start"); pre_full_gc_dump(gc_timer); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::clear(); #endif @@ -1395,7 +1395,7 @@ // not been removed from the discovered lists. ref_processor_stw()->enqueue_discovered_references(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::update_pointers(); #endif @@ -3607,7 +3607,7 @@ // FIXME: what is this about? // I'm ignoring the "fill_newgen()" call if "alloc_event_enabled" // is set. -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI assert(DerivedPointerTable::is_empty(), "derived pointer present"); #endif // always_do_update_barrier = true; @@ -3943,7 +3943,7 @@ verify_before_gc(); check_bitmaps("GC Start"); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::clear(); #endif @@ -5996,7 +5996,7 @@ enqueue_discovered_references(n_workers); redirty_logged_cards(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::update_pointers(); #endif }
--- a/src/share/vm/gc_implementation/g1/g1MarkSweep.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_implementation/g1/g1MarkSweep.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -89,7 +89,7 @@ mark_sweep_phase2(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // Don't add any more derived pointers during phase3 DerivedPointerTable::set_active(false); #endif @@ -172,7 +172,7 @@ if (VerifyDuringGC) { HandleMark hm; // handle scope -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTableDeactivate dpt_deact; #endif Universe::heap()->prepare_for_verify();
--- a/src/share/vm/gc_implementation/parallelScavenge/psMarkSweep.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_implementation/parallelScavenge/psMarkSweep.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -194,7 +194,7 @@ allocate_stacks(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::clear(); #endif @@ -205,7 +205,7 @@ mark_sweep_phase2(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // Don't add any more derived pointers during phase3 assert(DerivedPointerTable::is_active(), "Sanity"); DerivedPointerTable::set_active(false); @@ -258,7 +258,7 @@ CodeCache::gc_epilogue(); JvmtiExport::gc_epilogue(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::update_pointers(); #endif
--- a/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_implementation/parallelScavenge/psParallelCompact.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -1068,7 +1068,7 @@ CodeCache::gc_epilogue(); JvmtiExport::gc_epilogue(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::update_pointers(); #endif @@ -2070,7 +2070,7 @@ CodeCache::gc_prologue(); Threads::gc_prologue(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::clear(); #endif @@ -2086,7 +2086,7 @@ && gc_cause == GCCause::_java_lang_system_gc; summary_phase(vmthread_cm, maximum_heap_compaction || max_on_system_gc); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI assert(DerivedPointerTable::is_active(), "Sanity"); DerivedPointerTable::set_active(false); #endif
--- a/src/share/vm/gc_implementation/parallelScavenge/psScavenge.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_implementation/parallelScavenge/psScavenge.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -360,7 +360,7 @@ } save_to_space_top_before_gc(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::clear(); #endif @@ -635,7 +635,7 @@ assert(young_gen->to_space()->is_empty(), "to space should be empty now"); } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::update_pointers(); #endif
--- a/src/share/vm/gc_interface/collectedHeap.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/gc_interface/collectedHeap.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -217,7 +217,7 @@ #ifdef COMPILER2 _defer_initial_card_mark = ReduceInitialCardMarks && can_elide_tlab_store_barriers() && (DeferInitialCardMark || card_mark_must_follow_store()); -#elif defined JVMCI +#elif INCLUDE_JVMCI _defer_initial_card_mark = JVMCIDeferredInitBarriers && can_elide_tlab_store_barriers() && (DeferInitialCardMark || card_mark_must_follow_store()); #else @@ -527,7 +527,7 @@ " to threads list is doomed to failure!"); for (JavaThread *thread = Threads::first(); thread; thread = thread->next()) { if (use_tlab) thread->tlab().make_parsable(retire_tlabs); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // The deferred store barriers must all have been flushed to the // card-table (or other remembered set structure) before GC starts // processing the card-table (or other remembered set).
--- a/src/share/vm/interpreter/interpreter.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/interpreter/interpreter.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -409,7 +409,7 @@ address AbstractInterpreter::deopt_reexecute_entry(Method* method, address bcp) { assert(method->contains(bcp), "just checkin'"); Bytecodes::Code code = Bytecodes::java_code_at(method, bcp); -#if defined(COMPILER1) || defined(JVMCI) +#if defined(COMPILER1) || INCLUDE_JVMCI if(code == Bytecodes::_athrow ) { return Interpreter::rethrow_exception_entry(); }
--- a/src/share/vm/interpreter/interpreterRuntime.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/interpreter/interpreterRuntime.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -475,7 +475,7 @@ } } while (should_repeat == true); -#ifdef JVMCI +#if INCLUDE_JVMCI if (h_method->method_data() != NULL) { ResourceMark rm(thread); ProfileData* pdata = h_method->method_data()->allocate_bci_to_data(current_bci, NULL);
--- a/src/share/vm/interpreter/linkResolver.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/interpreter/linkResolver.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -125,11 +125,11 @@ private: static void lookup_method_in_klasses (methodHandle& result, KlassHandle klass, Symbol* name, Symbol* signature, bool checkpolymorphism, bool in_imethod_resolve, TRAPS); -#ifdef JVMCI +#if INCLUDE_JVMCI public: #endif static void lookup_instance_method_in_klasses (methodHandle& result, KlassHandle klass, Symbol* name, Symbol* signature, TRAPS); -#ifdef JVMCI +#if INCLUDE_JVMCI private: #endif static void lookup_method_in_interfaces (methodHandle& result, KlassHandle klass, Symbol* name, Symbol* signature, TRAPS); @@ -145,12 +145,12 @@ static void linktime_resolve_static_method (methodHandle& resolved_method, KlassHandle resolved_klass, Symbol* method_name, Symbol* method_signature, KlassHandle current_klass, bool check_access, TRAPS); static void linktime_resolve_special_method (methodHandle& resolved_method, KlassHandle resolved_klass, Symbol* method_name, Symbol* method_signature, KlassHandle current_klass, bool check_access, TRAPS); -#ifdef JVMCI +#if INCLUDE_JVMCI public: #endif static void linktime_resolve_virtual_method (methodHandle &resolved_method, KlassHandle resolved_klass, Symbol* method_name, Symbol* method_signature,KlassHandle current_klass, bool check_access, TRAPS); static void linktime_resolve_interface_method (methodHandle& resolved_method, KlassHandle resolved_klass, Symbol* method_name, Symbol* method_signature, KlassHandle current_klass, bool check_access, TRAPS); -#ifdef JVMCI +#if INCLUDE_JVMCI private: #endif
--- a/src/share/vm/memory/genCollectedHeap.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/memory/genCollectedHeap.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -452,7 +452,7 @@ } Universe::verify(" VerifyBeforeGC:"); } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::clear(); #endif @@ -514,7 +514,7 @@ } } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI DerivedPointerTable::update_pointers(); #endif @@ -1224,7 +1224,7 @@ }; void GenCollectedHeap::gc_epilogue(bool full) { -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI assert(DerivedPointerTable::is_empty(), "derived pointer present"); size_t actual_gap = pointer_delta((HeapWord*) (max_uintx-3), *(end_addr())); guarantee(actual_gap > (size_t)FastAllocateSizeLimit, "inline allocation wraps");
--- a/src/share/vm/memory/genMarkSweep.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/memory/genMarkSweep.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -96,7 +96,7 @@ mark_sweep_phase2(); // Don't add any more derived pointers during phase3 -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI assert(DerivedPointerTable::is_active(), "Sanity"); DerivedPointerTable::set_active(false); #endif
--- a/src/share/vm/memory/referenceProcessor.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/memory/referenceProcessor.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -57,7 +57,7 @@ java_lang_ref_SoftReference::set_clock(_soft_ref_timestamp_clock); _always_clear_soft_ref_policy = new AlwaysClearPolicy(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI _default_soft_ref_policy = new LRUMaxHeapPolicy(); #else _default_soft_ref_policy = new LRUCurrentHeapPolicy();
--- a/src/share/vm/oops/instanceKlass.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/oops/instanceKlass.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -61,7 +61,7 @@ #include "services/threadService.hpp" #include "utilities/dtrace.hpp" #include "utilities/macros.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "classfile/javaAssertions.hpp" #include "jvmci/jvmciRuntime.hpp" #endif
--- a/src/share/vm/oops/method.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/oops/method.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -218,7 +218,7 @@ Thread* myThread = Thread::current(); methodHandle h_this(myThread, this); -#if defined(ASSERT) && !defined(JVMCI) +#if defined(ASSERT) && !INCLUDE_JVMCI bool has_capability = myThread->is_VM_thread() || myThread->is_ConcurrentGC_thread() || myThread->is_GC_task_thread();
--- a/src/share/vm/oops/methodData.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/oops/methodData.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -420,7 +420,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI void VirtualCallData::clean_weak_klass_links(BoolObjectClosure* is_alive_cl) { ReceiverTypeData::clean_weak_klass_links(is_alive_cl); for (uint row = 0; row < method_row_limit(); row++) { @@ -440,7 +440,7 @@ } } } -#endif // JVMCI +#endif // INCLUDE_JVMCI #ifndef PRODUCT void ReceiverTypeData::print_receiver_data_on(outputStream* st) const { @@ -449,7 +449,7 @@ for (row = 0; row < row_limit(); row++) { if (receiver(row) != NULL) entries++; } -#ifdef JVMCI +#if INCLUDE_JVMCI st->print_cr("count(%u) nonprofiled_count(%u) entries(%u)", count(), nonprofiled_count(), entries); #else st->print_cr("count(%u) entries(%u)", count(), entries); @@ -473,7 +473,7 @@ print_receiver_data_on(st); } -#ifdef JVMCI +#if INCLUDE_JVMCI void VirtualCallData::print_method_data_on(outputStream* st) const { uint row; int entries = 0; @@ -501,7 +501,7 @@ void VirtualCallData::print_data_on(outputStream* st, const char* extra) const { print_shared(st, "VirtualCallData", extra); print_receiver_data_on(st); -#ifdef JVMCI +#if INCLUDE_JVMCI print_method_data_on(st); #endif } @@ -740,7 +740,7 @@ } int MethodData::bytecode_cell_count(Bytecodes::Code code) { -#if defined(COMPILER1) && !(defined(COMPILER2) || defined(JVMCI)) +#if defined(COMPILER1) && !(defined(COMPILER2) || INCLUDE_JVMCI) return no_profile_data; #else switch (code) { @@ -869,7 +869,7 @@ return false; } -#ifdef JVMCI +#if INCLUDE_JVMCI int MethodData::compute_extra_data_count(int data_size, int empty_bc_count, bool needs_speculative_traps) { if (!ProfileTraps) return 0; @@ -963,7 +963,7 @@ // the segment in bytes. int MethodData::initialize_data(BytecodeStream* stream, int data_index) { -#if defined(COMPILER1) && !(defined(COMPILER2) || defined(JVMCI)) +#if defined(COMPILER1) && !(defined(COMPILER2) || INCLUDE_JVMCI) return 0; #else int cell_count = -1; @@ -1236,7 +1236,7 @@ _num_blocks = 0; _would_profile = unknown; -#ifdef JVMCI +#if INCLUDE_JVMCI _jvmci_ir_size = 0; #endif
--- a/src/share/vm/oops/methodData.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/oops/methodData.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -543,7 +543,7 @@ // saw a null operand (cast/aastore/instanceof) null_seen_flag = DataLayout::first_flag + 0 -#ifdef JVMCI +#if INCLUDE_JVMCI // bytecode threw any exception , exception_seen_flag = null_seen_flag + 1 #endif @@ -570,7 +570,7 @@ bool null_seen() { return flag_at(null_seen_flag); } void set_null_seen() { set_flag_at(null_seen_flag); } -#ifdef JVMCI +#if INCLUDE_JVMCI // true if an exception was thrown at the specific BCI bool exception_seen() { return flag_at(exception_seen_flag); } void set_exception_seen() { set_flag_at(exception_seen_flag); } @@ -1179,7 +1179,7 @@ class ReceiverTypeData : public CounterData { protected: enum { -#ifdef JVMCI +#if INCLUDE_JVMCI // Description of the different counters // ReceiverTypeData for instanceof/checkcast/aastore: // C1/C2: count is incremented on type overflow and decremented for failed type checks @@ -1271,7 +1271,7 @@ set_count(0); set_receiver(row, NULL); set_receiver_count(row, 0); -#ifdef JVMCI +#if INCLUDE_JVMCI if (!this->is_VirtualCallData()) { // if this is a ReceiverTypeData for JVMCI, the nonprofiled_count // must also be reset (see "Description of the different counters" above) @@ -1287,7 +1287,7 @@ static ByteSize receiver_count_offset(uint row) { return cell_offset(receiver_count_cell_index(row)); } -#ifdef JVMCI +#if INCLUDE_JVMCI static ByteSize nonprofiled_receiver_count_offset() { return cell_offset(nonprofiled_count_off_set); } @@ -1386,7 +1386,7 @@ } #endif // CC_INTERP -#ifdef JVMCI +#if INCLUDE_JVMCI static ByteSize method_offset(uint row) { return cell_offset(method_cell_index(row)); } @@ -1442,7 +1442,7 @@ #endif #ifndef PRODUCT -#ifdef JVMCI +#if INCLUDE_JVMCI void print_method_data_on(outputStream* st) const; #endif void print_data_on(outputStream* st, const char* extra = NULL) const; @@ -2210,7 +2210,7 @@ enum WouldProfile {unknown, no_profile, profile}; WouldProfile _would_profile; -#ifdef JVMCI +#if INCLUDE_JVMCI // Support for HotSpotMethodData.setCompiledIRSize(int) int _jvmci_ir_size; #endif
--- a/src/share/vm/precompiled/precompiled.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/precompiled/precompiled.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -295,9 +295,9 @@ # include "c1/c1_ValueType.hpp" # include "c1/c1_globals.hpp" #endif // COMPILER1 -#ifdef JVMCI +#if INCLUDE_JVMCI # include "jvmci/jvmciGlobals.hpp" -#endif // JVMCI +#endif // INCLUDE_JVMCI #if INCLUDE_ALL_GCS # include "gc_implementation/concurrentMarkSweep/cmsOopClosures.hpp" # include "gc_implementation/concurrentMarkSweep/compactibleFreeListSpace.hpp"
--- a/src/share/vm/prims/jni.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/prims/jni.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -33,7 +33,7 @@ #include "classfile/vmSymbols.hpp" #include "interpreter/linkResolver.hpp" #include "utilities/macros.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciCompiler.hpp" #include "jvmci/jvmciRuntime.hpp" #endif @@ -5209,7 +5209,7 @@ *vm = (JavaVM *)(&main_vm); *(JNIEnv**)penv = thread->jni_environment(); -#if defined(JVMCI) +#if INCLUDE_JVMCI // We turn off CompileTheWorld so that compilation requests are not ignored during bootstrap or that JVMCI can be compiled by C1/C2. bool doCTW = CompileTheWorld; CompileTheWorld = false; @@ -5220,14 +5220,14 @@ if (FLAG_IS_DEFAULT(BootstrapJVMCI) ? !TieredCompilation : BootstrapJVMCI) { JVMCICompiler::instance()->bootstrap(); } -#elif defined(JVMCI) +#elif INCLUDE_JVMCI if (doCTW) { // required for hosted CTW. CompilationPolicy::completed_vm_startup(); } #endif -#if defined(JVMCI) +#if INCLUDE_JVMCI if (doCTW) { JVMCICompiler::instance()->compile_the_world(); }
--- a/src/share/vm/prims/jvm.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/prims/jvm.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -383,7 +383,7 @@ const char* compiler_name = "HotSpot " CSIZE "Client Compiler"; #elif defined(COMPILER2) const char* compiler_name = "HotSpot " CSIZE "Server Compiler"; -#elif defined(JVMCI) +#elif INCLUDE_JVMCI const char* compiler_name = "HotSpot " CSIZE "JVMCI Compiler"; #else const char* compiler_name = "";
--- a/src/share/vm/prims/nativeLookup.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/prims/nativeLookup.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -126,7 +126,7 @@ void JNICALL JVM_RegisterMethodHandleMethods(JNIEnv *env, jclass unsafecls); void JNICALL JVM_RegisterPerfMethods(JNIEnv *env, jclass perfclass); void JNICALL JVM_RegisterWhiteBoxMethods(JNIEnv *env, jclass wbclass); -#ifdef JVMCI +#if INCLUDE_JVMCI void JNICALL JVM_InitJVMCIClassLoader(JNIEnv *env, jclass c, jobject loader); void JNICALL JVM_InitializeJVMCINatives(JNIEnv *env, jclass compilerToVMClass); jobject JNICALL JVM_GetJVMCIRuntime(JNIEnv *env, jclass c); @@ -143,7 +143,7 @@ { CC"Java_java_lang_invoke_MethodHandleNatives_registerNatives", NULL, FN_PTR(JVM_RegisterMethodHandleMethods) }, { CC"Java_sun_misc_Perf_registerNatives", NULL, FN_PTR(JVM_RegisterPerfMethods) }, { CC"Java_sun_hotspot_WhiteBox_registerNatives", NULL, FN_PTR(JVM_RegisterWhiteBoxMethods) }, -#ifdef JVMCI +#if INCLUDE_JVMCI { CC"Java_jdk_internal_jvmci_service_JVMCIClassLoaderFactory_init", NULL, FN_PTR(JVM_InitJVMCIClassLoader) }, { CC"Java_jdk_internal_jvmci_runtime_JVMCI_initializeRuntime", NULL, FN_PTR(JVM_GetJVMCIRuntime) }, { CC"Java_jdk_internal_jvmci_service_Services_getServiceImpls", NULL, FN_PTR(JVM_GetJVMCIServiceImpls) },
--- a/src/share/vm/runtime/arguments.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/arguments.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -44,7 +44,7 @@ #include "utilities/macros.hpp" #include "utilities/stringUtils.hpp" #include "utilities/taskqueue.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciRuntime.hpp" #endif #ifdef TARGET_OS_FAMILY_linux @@ -98,7 +98,7 @@ int Arguments::_num_jvm_flags = 0; char** Arguments::_jvm_args_array = NULL; int Arguments::_num_jvm_args = 0; -#ifdef JVMCI +#if INCLUDE_JVMCI char** Arguments::_jvmci_args_array = NULL; int Arguments::_num_jvmci_args = 0; #endif @@ -217,7 +217,7 @@ // Set OS specific system properties values os::init_system_properties_values(); -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCIRuntime::parse_properties(&_system_properties); #endif } @@ -817,7 +817,7 @@ void Arguments::build_jvm_flags(const char* arg) { add_string(&_jvm_flags_array, &_num_jvm_flags, arg); } -#ifdef JVMCI +#if INCLUDE_JVMCI void Arguments::add_jvmci_arg(const char* arg) { add_string(&_jvmci_args_array, &_num_jvmci_args, arg); } @@ -1127,7 +1127,7 @@ } } -#if defined(COMPILER2) || defined(JVMCI) || defined(_LP64) || !INCLUDE_CDS +#if defined(COMPILER2) || INCLUDE_JVMCI || defined(_LP64) || !INCLUDE_CDS // Conflict: required to use shared spaces (-Xshare:on), but // incompatible command line options were chosen. @@ -1595,7 +1595,7 @@ void Arguments::set_ergonomics_flags() { select_gc(); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // Shared spaces work fine with other GCs but causes bytecode rewriting // to be disabled, which hurts interpreter performance and decreases // server performance. When -server is specified, keep the default off @@ -1679,7 +1679,7 @@ void Arguments::set_g1_gc_flags() { assert(UseG1GC, "Error"); -#if defined(COMPILER1) || defined(JVMCI) +#if defined(COMPILER1) || INCLUDE_JVMCI FastTLABRefill = false; #endif FLAG_SET_DEFAULT(ParallelGCThreads, @@ -2446,7 +2446,7 @@ } #endif } -#ifdef JVMCI +#if INCLUDE_JVMCI if (UseG1GC) { if (IgnoreUnrecognizedVMOptions) { FLAG_SET_CMDLINE(bool, UseG1GC, true); @@ -2783,7 +2783,7 @@ return JNI_ERR; } #endif // !INCLUDE_JVMTI -#if defined(JVMCI) +#if INCLUDE_JVMCI if (strcmp(name, "hprof") == 0) { FLAG_SET_CMDLINE(bool, JVMCIHProfEnabled, true); } @@ -2810,7 +2810,7 @@ return JNI_ERR; } #endif // !INCLUDE_JVMTI -#if defined(JVMCI) +#if INCLUDE_JVMCI if (valid_hprof_or_jdwp_agent(name, is_absolute_path)) { FLAG_SET_CMDLINE(bool, JVMCIHProfEnabled, true); } @@ -3389,7 +3389,7 @@ } } } -#ifdef JVMCI +#if INCLUDE_JVMCI else if (match_option(option, "-G:", &tail)) { // -G:XXX // Option for the JVMCI compiler. if (PrintVMOptions) { @@ -3615,7 +3615,7 @@ // This must be done after all -D arguments have been processed. scp_p->expand_endorsed(); -#ifdef JVMCI +#if INCLUDE_JVMCI if (!UseJVMCIClassLoader) { // Append lib/jvmci/*.jar to boot class path char jvmciDir[JVM_MAXPATHLEN]; @@ -4136,9 +4136,9 @@ #ifdef COMPILER1 || !UseFastLocking #endif // COMPILER1 -#ifdef JVMCI +#if INCLUDE_JVMCI || !JVMCIUseFastLocking -#endif // JVMCI +#endif ) { if (!FLAG_IS_DEFAULT(UseBiasedLocking) && UseBiasedLocking) { // flag set to true on command line; warn the user that they
--- a/src/share/vm/runtime/arguments.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/arguments.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -246,7 +246,7 @@ // an array containing all jvm arguments specified in the command line static char** _jvm_args_array; static int _num_jvm_args; -#ifdef JVMCI +#if INCLUDE_JVMCI // an array containing all JVMCI arguments specified in the command line static char** _jvmci_args_array; static int _num_jvmci_args; @@ -410,7 +410,7 @@ // methods to build strings from individual args static void build_jvm_args(const char* arg); static void build_jvm_flags(const char* arg); -#ifdef JVMCI +#if INCLUDE_JVMCI static void add_jvmci_arg(const char* arg); #endif static void add_string(char*** bldarray, int* count, const char* arg); @@ -492,7 +492,7 @@ // return a char* array containing all options static char** jvm_flags_array() { return _jvm_flags_array; } static char** jvm_args_array() { return _jvm_args_array; } -#ifdef JVMCI +#if INCLUDE_JVMCI static char** jvmci_args_array() { return _jvmci_args_array; } static int num_jvmci_args() { return _num_jvmci_args; } #endif
--- a/src/share/vm/runtime/deoptimization.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/deoptimization.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -92,7 +92,7 @@ PRAGMA_FORMAT_MUTE_WARNINGS_FOR_GCC -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciRuntime.hpp" #include "jvmci/jvmciJavaAccess.hpp" #endif @@ -225,7 +225,7 @@ bool realloc_failures = false; -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // Reallocate the non-escaping objects and restore their fields. Then // relock objects if synchronization on them was eliminated. #ifdef COMPILER2 @@ -313,7 +313,7 @@ } } #endif // COMPILER2 -#endif // COMPILER2 || JVMCI +#endif // COMPILER2 || INCLUDE_JVMCI // Ensure that no safepoint is taken after pointers have been stored // in fields of rematerialized objects. If a safepoint occurs from here on @@ -321,7 +321,7 @@ No_Safepoint_Verifier no_safepoint; vframeArray* array = create_vframeArray(thread, deoptee, &map, chunk, realloc_failures); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI if (realloc_failures) { pop_frames_failed_reallocs(thread, array); } @@ -776,7 +776,7 @@ } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI bool Deoptimization::realloc_objects(JavaThread* thread, frame* fr, GrowableArray<ScopeValue*>* objects, TRAPS) { Handle pending_exception(thread->pending_exception()); const char* exception_file = thread->exception_file(); @@ -1111,7 +1111,7 @@ } } #endif -#endif // COMPILER2 || JVMCI +#endif // COMPILER2 || INCLUDE_JVMCI vframeArray* Deoptimization::create_vframeArray(JavaThread* thread, frame fr, RegisterMap *reg_map, GrowableArray<compiledVFrame*>* chunk, bool realloc_failures) { Events::log(thread, "DEOPT PACKING pc=" INTPTR_FORMAT " sp=" INTPTR_FORMAT, fr.pc(), fr.sp()); @@ -1171,7 +1171,7 @@ return array; } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI void Deoptimization::pop_frames_failed_reallocs(JavaThread* thread, vframeArray* array) { // Reallocation of some scalar replaced objects failed. Record // that we need to pop all the interpreter frames for the @@ -1361,7 +1361,7 @@ JRT_END -#if defined(COMPILER2) || defined(SHARK) || defined(JVMCI) +#if defined(COMPILER2) || defined(SHARK) || INCLUDE_JVMCI void Deoptimization::load_class_by_index(constantPoolHandle constant_pool, int index, TRAPS) { // in case of an unresolved klass entry, load the class. if (constant_pool->tag_at(index).is_unresolved_klass()) { @@ -1424,7 +1424,7 @@ thread->inc_in_deopt_handler(); // We need to update the map if we have biased locking. -#ifdef JVMCI +#if INCLUDE_JVMCI // (lstadler) JVMCI might need to get an exception from the stack, which in turn requires the register map to be valid RegisterMap reg_map(thread, true); #else @@ -1448,7 +1448,7 @@ DeoptReason reason = trap_request_reason(trap_request); DeoptAction action = trap_request_action(trap_request); -#ifdef JVMCI +#if INCLUDE_JVMCI int debug_id = trap_request_debug_id(trap_request); #endif jint unloaded_class_index = trap_request_index(trap_request); // CP idx or -1 @@ -1463,7 +1463,7 @@ if (TraceDeoptimization) { ttyLocker ttyl; tty->print_cr(" bci=%d pc=" INTPTR_FORMAT ", relative_pc=%d, method=%s" JVMCI_ONLY(", debug_id=%d"), trap_scope->bci(), fr.pc(), fr.pc() - nm->code_begin(), trap_scope->method()->name_and_sig_as_C_string() -#ifdef JVMCI +#if INCLUDE_JVMCI , debug_id #endif ); @@ -1471,7 +1471,7 @@ methodHandle trap_method = trap_scope->method(); int trap_bci = trap_scope->bci(); -#ifdef JVMCI +#if INCLUDE_JVMCI oop speculation = thread->pending_failed_speculation(); if (nm->is_compiled_by_jvmci()) { if (speculation != NULL) { @@ -1534,7 +1534,7 @@ bool create_if_missing = ProfileTraps RTM_OPT_ONLY( || UseRTMLocking ); methodHandle profiled_method; -#ifdef JVMCI +#if INCLUDE_JVMCI if (nm->is_compiled_by_jvmci()) { profiled_method = nm->method(); } else { @@ -1604,7 +1604,7 @@ tty->print("Uncommon trap occurred in"); nm->method()->print_short_name(tty); tty->print(" compiler=%s compile_id=%d", nm->compiler() == NULL ? "" : nm->compiler()->name(), nm->compile_id()); -#ifdef JVMCI +#if INCLUDE_JVMCI oop installedCode = nm->jvmci_installed_code(); if (installedCode != NULL) { oop installedCodeName = NULL; @@ -1619,14 +1619,14 @@ } else if (nm->is_compiled_by_jvmci()) { tty->print(" (JVMCI: no installed code) "); } -#endif //JVMCI +#endif tty->print(" (@" INTPTR_FORMAT ") thread=" UINTX_FORMAT " reason=%s action=%s unloaded_class_index=%d" JVMCI_ONLY(" debug_id=%d"), fr.pc(), os::current_thread_id(), trap_reason_name(reason), trap_action_name(action), unloaded_class_index -#ifdef JVMCI +#if INCLUDE_JVMCI , debug_id #endif ); @@ -1764,7 +1764,7 @@ bool maybe_prior_trap = false; bool maybe_prior_recompile = false; pdata = query_update_method_data(trap_mdo, trap_bci, reason, true, -#ifdef JVMCI +#if INCLUDE_JVMCI nm->is_compiled_by_jvmci() && nm->is_osr_method(), #endif nm->method(), @@ -1913,7 +1913,7 @@ int trap_bci, Deoptimization::DeoptReason reason, bool update_total_trap_count, -#ifdef JVMCI +#if INCLUDE_JVMCI bool is_osr, #endif Method* compiled_method, @@ -1926,7 +1926,7 @@ uint this_trap_count = 0; if (update_total_trap_count) { uint idx = reason; -#ifdef JVMCI +#if INCLUDE_JVMCI if (is_osr) { idx += Reason_LIMIT; } @@ -2005,7 +2005,7 @@ query_update_method_data(trap_mdo, trap_bci, (DeoptReason)reason, update_total_counts, -#ifdef JVMCI +#if INCLUDE_JVMCI false, #endif NULL, @@ -2142,7 +2142,7 @@ "speculate_class_check", "rtm_state_change", "unstable_if" -#ifdef JVMCI +#if INCLUDE_JVMCI "aliasing", "transfer_to_interpreter", "not_compiled_exception_handler", @@ -2181,21 +2181,21 @@ jint unloaded_class_index = trap_request_index(trap_request); const char* reason = trap_reason_name(trap_request_reason(trap_request)); const char* action = trap_action_name(trap_request_action(trap_request)); -#ifdef JVMCI +#if INCLUDE_JVMCI int debug_id = trap_request_debug_id(trap_request); #endif size_t len; if (unloaded_class_index < 0) { len = jio_snprintf(buf, buflen, "reason='%s' action='%s'" JVMCI_ONLY(" debug_id='%d'"), reason, action -#ifdef JVMCI +#if INCLUDE_JVMCI ,debug_id #endif ); } else { len = jio_snprintf(buf, buflen, "reason='%s' action='%s' index='%d'" JVMCI_ONLY(" debug_id='%d'"), reason, action, unloaded_class_index -#ifdef JVMCI +#if INCLUDE_JVMCI ,debug_id #endif ); @@ -2295,7 +2295,7 @@ if (xtty != NULL) xtty->tail("statistics"); } } -#else // COMPILER2 || SHARK || JVMCI +#else // COMPILER2 || SHARK || INCLUDE_JVMCI // Stubs for C1 only system. @@ -2331,4 +2331,4 @@ return buf; } -#endif // COMPILER2 || SHARK || JVMCI +#endif // COMPILER2 || SHARK || INCLUDE_JVMCI
--- a/src/share/vm/runtime/deoptimization.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/deoptimization.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -56,7 +56,7 @@ Reason_intrinsic, // saw unexpected operand to intrinsic (@bci) Reason_bimorphic, // saw unexpected object class in bimorphic inlining (@bci) -#ifdef JVMCI +#if INCLUDE_JVMCI Reason_unreached0 = Reason_null_assert, Reason_type_checked_inlining = Reason_intrinsic, Reason_optimized_type_check = Reason_bimorphic, @@ -75,7 +75,7 @@ Reason_speculate_class_check, // saw unexpected object class from type speculation Reason_rtm_state_change, // rtm state change detected Reason_unstable_if, // a branch predicted always false was taken -#ifdef JVMCI +#if INCLUDE_JVMCI Reason_aliasing, // optimistic assumption about aliasing failed Reason_transfer_to_interpreter, // explicit transferToInterpreter() Reason_not_compiled_exception_handler, @@ -138,7 +138,7 @@ // executing in a particular CodeBlob if UseBiasedLocking is enabled static void revoke_biases_of_monitors(CodeBlob* cb); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI JVMCI_ONLY(public:) // Support for restoring non-escaping objects @@ -149,7 +149,7 @@ static void relock_objects(GrowableArray<MonitorInfo*>* monitors, JavaThread* thread, bool realloc_failures); static void pop_frames_failed_reallocs(JavaThread* thread, vframeArray* array); NOT_PRODUCT(static void print_objects(GrowableArray<ScopeValue*>* objects, bool realloc_failures);) -#endif // COMPILER2 || JVMCI +#endif // COMPILER2 || INCLUDE_JVMCI public: static vframeArray* create_vframeArray(JavaThread* thread, frame fr, RegisterMap *reg_map, GrowableArray<compiledVFrame*>* chunk, bool realloc_failures); @@ -395,7 +395,7 @@ int trap_bci, DeoptReason reason, bool update_total_trap_count, -#ifdef JVMCI +#if INCLUDE_JVMCI bool is_osr, #endif Method* compiled_method,
--- a/src/share/vm/runtime/frame.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/frame.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -734,7 +734,7 @@ nm->compile_id(), (nm->is_osr_method() ? "%" : ""), ((nm->compiler() != NULL) ? nm->compiler()->name() : ""), buf, m->code_size(), _pc, _cb->code_begin(), _pc - _cb->code_begin()); -#ifdef JVMCI +#if INCLUDE_JVMCI char* jvmciName = nm->jvmci_installed_code_name(buf, buflen); if (jvmciName != NULL) { st->print(" (%s)", jvmciName); @@ -1313,7 +1313,7 @@ // make sure we have the right receiver type } } -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI assert(DerivedPointerTable::is_empty(), "must be empty before verify"); #endif oops_do_internal(&VerifyOopClosure::verify_oop, NULL, NULL, (RegisterMap*)map, false);
--- a/src/share/vm/runtime/globals.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/globals.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -38,7 +38,7 @@ #ifdef COMPILER1 #include "c1/c1_globals.hpp" #endif -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciGlobals.hpp" #endif #ifdef COMPILER2 @@ -494,7 +494,7 @@ #ifdef COMPILER1 C1_FLAGS(C1_DEVELOP_FLAG_STRUCT, C1_PD_DEVELOP_FLAG_STRUCT, C1_PRODUCT_FLAG_STRUCT, C1_PD_PRODUCT_FLAG_STRUCT, C1_DIAGNOSTIC_FLAG_STRUCT, C1_NOTPRODUCT_FLAG_STRUCT) #endif -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCI_FLAGS(JVMCI_DEVELOP_FLAG_STRUCT, JVMCI_PD_DEVELOP_FLAG_STRUCT, JVMCI_PRODUCT_FLAG_STRUCT, JVMCI_PD_PRODUCT_FLAG_STRUCT, JVMCI_NOTPRODUCT_FLAG_STRUCT) #endif #ifdef COMPILER2
--- a/src/share/vm/runtime/globals_extension.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/globals_extension.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -86,7 +86,7 @@ #ifdef COMPILER1 C1_FLAGS(C1_DEVELOP_FLAG_MEMBER, C1_PD_DEVELOP_FLAG_MEMBER, C1_PRODUCT_FLAG_MEMBER, C1_PD_PRODUCT_FLAG_MEMBER, C1_DIAGNOSTIC_FLAG_MEMBER, C1_NOTPRODUCT_FLAG_MEMBER) #endif -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCI_FLAGS(JVMCI_DEVELOP_FLAG_MEMBER, JVMCI_PD_DEVELOP_FLAG_MEMBER, JVMCI_PRODUCT_FLAG_MEMBER, JVMCI_PD_PRODUCT_FLAG_MEMBER, JVMCI_NOTPRODUCT_FLAG_MEMBER) #endif #ifdef COMPILER2 @@ -180,7 +180,7 @@ C1_DIAGNOSTIC_FLAG_MEMBER_WITH_TYPE, C1_NOTPRODUCT_FLAG_MEMBER_WITH_TYPE) #endif -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCI_FLAGS(JVMCI_DEVELOP_FLAG_MEMBER_WITH_TYPE, JVMCI_PD_DEVELOP_FLAG_MEMBER_WITH_TYPE, JVMCI_PRODUCT_FLAG_MEMBER_WITH_TYPE,
--- a/src/share/vm/runtime/java.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/java.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -30,7 +30,7 @@ #include "compiler/compileBroker.hpp" #include "compiler/compilerOracle.hpp" #include "interpreter/bytecodeHistogram.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciCompiler.hpp" #include "jvmci/jvmciRuntime.hpp" #endif @@ -490,7 +490,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCIRuntime::shutdown(); #endif
--- a/src/share/vm/runtime/javaCalls.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/javaCalls.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -40,7 +40,7 @@ #include "runtime/signature.hpp" #include "runtime/stubRoutines.hpp" #include "runtime/thread.inline.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciJavaAccess.hpp" #include "jvmci/jvmciRuntime.hpp" #endif @@ -343,7 +343,7 @@ CHECK_UNHANDLED_OOPS_ONLY(thread->clear_unhandled_oops();) -#ifdef JVMCI +#if INCLUDE_JVMCI nmethod* nm = args->alternative_target(); if (nm == NULL) { #endif @@ -353,7 +353,7 @@ args->verify(method, result->get_type(), thread); } else debug_only(args->verify(method, result->get_type(), thread)); -#ifdef JVMCI +#if INCLUDE_JVMCI } #else @@ -418,7 +418,7 @@ os::bang_stack_shadow_pages(); } -#ifdef JVMCI +#if INCLUDE_JVMCI if (nm != NULL) { if (nm->is_alive()) { ((JavaThread*) THREAD)->set_jvmci_alternate_call_target(nm->verified_entry_point());
--- a/src/share/vm/runtime/javaCalls.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/javaCalls.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -100,7 +100,7 @@ int _size; int _max_size; bool _start_at_zero; // Support late setting of receiver -#ifdef JVMCI +#if INCLUDE_JVMCI nmethod* _alternative_target; // Nmethod that should be called instead of normal target #endif @@ -140,7 +140,7 @@ } } -#ifdef JVMCI +#if INCLUDE_JVMCI void set_alternative_target(nmethod* target) { _alternative_target = target; }
--- a/src/share/vm/runtime/os.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/os.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -1189,7 +1189,7 @@ "%/lib/jce.jar:" "%/lib/charsets.jar:" "%/lib/jfr.jar:" -#ifdef JVMCI +#if INCLUDE_JVMCI "%/lib/jvmci-service.jar:" #endif "%/classes";
--- a/src/share/vm/runtime/os.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/os.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -104,7 +104,7 @@ class os: AllStatic { friend class VMStructs; friend class MallocTracker; -#ifdef JVMCI +#if INCLUDE_JVMCI friend class Arguments; // need access to format_boot_path #endif
--- a/src/share/vm/runtime/rframe.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/rframe.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -152,7 +152,7 @@ void RFrame::print(const char* kind) { #ifndef PRODUCT -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI int cnt = top_method()->interpreter_invocation_count(); #else int cnt = top_method()->invocation_count();
--- a/src/share/vm/runtime/sharedRuntime.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/sharedRuntime.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -111,7 +111,7 @@ _resolve_virtual_call_blob = generate_resolve_blob(CAST_FROM_FN_PTR(address, SharedRuntime::resolve_virtual_call_C), "resolve_virtual_call"); _resolve_static_call_blob = generate_resolve_blob(CAST_FROM_FN_PTR(address, SharedRuntime::resolve_static_call_C), "resolve_static_call"); -#if defined(COMPILER2) || defined(JVMCI) +#if defined(COMPILER2) || INCLUDE_JVMCI // Vectors are generated only by C2 and JVMCI. bool support_wide = is_wide_vector(MaxVectorSize); if (support_wide) { @@ -488,7 +488,7 @@ // Reset method handle flag. thread->set_is_method_handle_return(false); -#ifdef JVMCI +#if INCLUDE_JVMCI // JVMCI's ExceptionHandlerStub expects the thread local exception PC to be clear // and other exception handler continuations do not read it thread->set_exception_pc(NULL); @@ -650,7 +650,7 @@ assert(nm != NULL, "must exist"); ResourceMark rm; -#ifdef JVMCI +#if INCLUDE_JVMCI if (nm->is_compiled_by_jvmci()) { // lookup exception handler for this pc int catch_pco = ret_pc - nm->code_begin(); @@ -795,7 +795,7 @@ throw_and_post_jvmti_exception(thread, exception); JRT_END -#ifdef JVMCI +#if INCLUDE_JVMCI address SharedRuntime::deoptimize_for_implicit_exception(JavaThread* thread, address pc, nmethod* nm, int deopt_reason) { assert(deopt_reason > Deoptimization::Reason_none && deopt_reason < Deoptimization::Reason_LIMIT, "invalid deopt reason"); thread->set_jvmci_implicit_exception_pc(pc); @@ -901,7 +901,7 @@ #ifndef PRODUCT _implicit_null_throws++; #endif -#ifdef JVMCI +#if INCLUDE_JVMCI if (nm->is_compiled_by_jvmci() && nm->pc_desc_at(pc) != NULL) { // If there's no PcDesc then we'll die way down inside of // deopt instead of just getting normal error reporting, @@ -910,7 +910,7 @@ } else { #endif target_pc = nm->continuation_for_implicit_exception(pc); -#ifdef JVMCI +#if INCLUDE_JVMCI } #endif // If there's an unexpected fault, target_pc might be NULL, @@ -928,13 +928,13 @@ #ifndef PRODUCT _implicit_div0_throws++; #endif -#ifdef JVMCI +#if INCLUDE_JVMCI if (nm->is_compiled_by_jvmci() && nm->pc_desc_at(pc) != NULL) { return deoptimize_for_implicit_exception(thread, pc, nm, Deoptimization::Reason_div0_check); } else { #endif target_pc = nm->continuation_for_implicit_exception(pc); -#ifdef JVMCI +#if INCLUDE_JVMCI } #endif // If there's an unexpected fault, target_pc might be NULL, @@ -1013,7 +1013,7 @@ JRT_ENTRY_NO_ASYNC(void, SharedRuntime::register_finalizer(JavaThread* thread, oopDesc* obj)) -#ifdef JVMCI +#if INCLUDE_JVMCI if (!obj->klass()->has_finalizer()) { return; }
--- a/src/share/vm/runtime/sharedRuntime.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/sharedRuntime.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -193,7 +193,7 @@ static address continuation_for_implicit_exception(JavaThread* thread, address faulting_pc, ImplicitExceptionKind exception_kind); -#ifdef JVMCI +#if INCLUDE_JVMCI static address deoptimize_for_implicit_exception(JavaThread* thread, address pc, nmethod* nm, int deopt_reason); #endif
--- a/src/share/vm/runtime/sweeper.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/sweeper.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -458,14 +458,14 @@ class NMethodMarker: public StackObj { private: -#ifdef JVMCI +#if INCLUDE_JVMCI JavaThread* _thread; #else CompilerThread* _thread; #endif public: NMethodMarker(nmethod* nm) { -#ifdef JVMCI +#if INCLUDE_JVMCI _thread = JavaThread::current(); #else _thread = CompilerThread::current();
--- a/src/share/vm/runtime/thread.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/thread.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -29,7 +29,7 @@ #include "classfile/vmSymbols.hpp" #include "code/scopeDesc.hpp" #include "compiler/compileBroker.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "jvmci/jvmciCompiler.hpp" #include "jvmci/jvmciRuntime.hpp" #endif @@ -843,7 +843,7 @@ active_handles()->oops_do(f); // Do oop for ThreadShadow f->do_oop((oop*)&_pending_exception); -#ifdef JVMCI +#if INCLUDE_JVMCI f->do_oop((oop*)&_pending_failed_speculation); #endif handle_area()->oops_do(f); @@ -1426,7 +1426,7 @@ // ======= JavaThread ======== -#ifdef JVMCI +#if INCLUDE_JVMCI jlong* JavaThread::_jvmci_old_thread_counters; @@ -1451,7 +1451,7 @@ } } -#endif // JVMCI +#endif // A JavaThread is a normal Java thread @@ -1486,7 +1486,7 @@ _in_deopt_handler = 0; _doing_unsafe_access = false; _stack_guard_state = stack_guard_unused; -#ifdef JVMCI +#if INCLUDE_JVMCI _jvmci_alternate_call_target = NULL; _jvmci_implicit_exception_pc = NULL; if (JVMCICounterSize > 0) { @@ -1495,7 +1495,7 @@ } else { _jvmci_counters = NULL; } -#endif // JVMCI +#endif (void)const_cast<oop&>(_exception_oop = NULL); _exception_pc = 0; _exception_handler_pc = 0; @@ -1676,7 +1676,7 @@ if (_thread_profiler != NULL) delete _thread_profiler; if (_thread_stat != NULL) delete _thread_stat; -#ifdef JVMCI +#if INCLUDE_JVMCI if (JVMCICounterSize > 0) { if (jvmci_counters_include(this)) { for (int i = 0; i < JVMCICounterSize; i++) { @@ -1685,7 +1685,7 @@ } FREE_C_HEAP_ARRAY(jlong, _jvmci_counters, mtInternal); } -#endif // JVMCI +#endif } @@ -3395,7 +3395,7 @@ jint parse_result = Arguments::parse(args); if (parse_result != JNI_OK) return parse_result; -#ifdef JVMCI +#if INCLUDE_JVMCI OptionValuesTable* options = JVMCIRuntime::parse_arguments(); if (options == NULL) { return JNI_ERR; @@ -3456,14 +3456,14 @@ // Initialize global data structures and create system classes in heap vm_init_globals(); -#ifdef JVMCI +#if INCLUDE_JVMCI if (JVMCICounterSize > 0) { JavaThread::_jvmci_old_thread_counters = NEW_C_HEAP_ARRAY(jlong, JVMCICounterSize, mtInternal); memset(JavaThread::_jvmci_old_thread_counters, 0, sizeof(jlong) * JVMCICounterSize); } else { JavaThread::_jvmci_old_thread_counters = NULL; } -#endif // JVMCI +#endif // Attach the main thread to this os thread JavaThread* main_thread = new JavaThread(); @@ -3709,7 +3709,7 @@ Chunk::start_chunk_pool_cleaner_task(); } -#ifdef JVMCI +#if INCLUDE_JVMCI JVMCIRuntime::set_options(options, main_thread); delete options; #endif @@ -4133,11 +4133,11 @@ delete thread; -#ifdef JVMCI +#if INCLUDE_JVMCI if (JVMCICounterSize > 0) { FREE_C_HEAP_ARRAY(jlong, JavaThread::_jvmci_old_thread_counters, mtInternal); } -#endif // JVMCI +#endif // exit_globals() will delete tty exit_globals();
--- a/src/share/vm/runtime/thread.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/thread.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -916,7 +916,7 @@ private: -#ifdef JVMCI +#if INCLUDE_JVMCI address _jvmci_alternate_call_target; address _jvmci_implicit_exception_pc; // pc at which the most recent implicit exception occurred @@ -926,7 +926,7 @@ static jlong* _jvmci_old_thread_counters; static void collect_counters(typeArrayOop array); private: -#endif // JVMCI +#endif nmethod* _scanned_nmethod; // nmethod being scanned by the sweeper @@ -1301,7 +1301,7 @@ MemRegion deferred_card_mark() const { return _deferred_card_mark; } void set_deferred_card_mark(MemRegion mr) { _deferred_card_mark = mr; } -#ifdef JVMCI +#if INCLUDE_JVMCI void set_jvmci_alternate_call_target(address a) { _jvmci_alternate_call_target = a; } void set_jvmci_implicit_exception_pc(address a) { _jvmci_implicit_exception_pc = a; } #endif @@ -1399,11 +1399,11 @@ static ByteSize thread_state_offset() { return byte_offset_of(JavaThread, _thread_state ); } static ByteSize saved_exception_pc_offset() { return byte_offset_of(JavaThread, _saved_exception_pc ); } static ByteSize osthread_offset() { return byte_offset_of(JavaThread, _osthread ); } -#ifdef JVMCI +#if INCLUDE_JVMCI static ByteSize jvmci_alternate_call_target_offset() { return byte_offset_of(JavaThread, _jvmci_alternate_call_target); } static ByteSize jvmci_implicit_exception_pc_offset() { return byte_offset_of(JavaThread, _jvmci_implicit_exception_pc); } static ByteSize jvmci_counters_offset() { return byte_offset_of(JavaThread, _jvmci_counters ); } -#endif // JVMCI +#endif static ByteSize exception_oop_offset() { return byte_offset_of(JavaThread, _exception_oop ); } static ByteSize exception_pc_offset() { return byte_offset_of(JavaThread, _exception_pc ); } static ByteSize exception_handler_pc_offset() { return byte_offset_of(JavaThread, _exception_handler_pc); }
--- a/src/share/vm/runtime/vmStructs.cpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/vmStructs.cpp Wed Jul 22 08:56:03 2015 -0700 @@ -70,7 +70,7 @@ #include "oops/constMethod.hpp" #include "oops/constantPool.hpp" #include "oops/cpCache.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI #include "oops/fieldStreams.hpp" #endif #include "oops/instanceClassLoaderKlass.hpp" @@ -106,7 +106,7 @@ #include "utilities/globalDefinitions.hpp" #include "utilities/hashtable.hpp" #include "utilities/macros.hpp" -#ifdef JVMCI +#if INCLUDE_JVMCI # include "jvmci/jvmciRuntime.hpp" # include "jvmci/vmStructs_jvmci.hpp" #endif @@ -3048,7 +3048,7 @@ GENERATE_C1_UNCHECKED_STATIC_VM_STRUCT_ENTRY, GENERATE_C2_UNCHECKED_STATIC_VM_STRUCT_ENTRY) -#ifdef JVMCI +#if INCLUDE_JVMCI VM_STRUCTS_JVMCI(GENERATE_NONSTATIC_VM_STRUCT_ENTRY, GENERATE_STATIC_VM_STRUCT_ENTRY) #endif @@ -3097,7 +3097,7 @@ GENERATE_C2_VM_TYPE_ENTRY, GENERATE_C2_TOPLEVEL_VM_TYPE_ENTRY) -#ifdef JVMCI +#if INCLUDE_JVMCI VM_TYPES_JVMCI(GENERATE_VM_TYPE_ENTRY, GENERATE_TOPLEVEL_VM_TYPE_ENTRY) #endif @@ -3144,7 +3144,7 @@ GENERATE_C2_VM_INT_CONSTANT_ENTRY, GENERATE_C2_PREPROCESSOR_VM_INT_CONSTANT_ENTRY) -#ifdef JVMCI +#if INCLUDE_JVMCI VM_INT_CONSTANTS_JVMCI(GENERATE_VM_INT_CONSTANT_ENTRY, GENERATE_PREPROCESSOR_VM_INT_CONSTANT_ENTRY) @@ -3496,7 +3496,7 @@ #endif -#ifdef JVMCI +#if INCLUDE_JVMCI // Emit intialization code for HotSpotVMConfig. It's placed here so // it can take advantage of the relaxed access checking enjoyed by // VMStructs.
--- a/src/share/vm/runtime/vmStructs.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/runtime/vmStructs.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -126,7 +126,7 @@ static void test(); #endif -#ifdef JVMCI +#if INCLUDE_JVMCI // The definition of this method is generated by // jdk.internal.jvmci.hotspotvmconfig.HotSpotVMConfigProcessor. static void initHotSpotVMConfig(oop config);
--- a/src/share/vm/utilities/exceptions.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/utilities/exceptions.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -61,7 +61,7 @@ friend class VMStructs; protected: -#ifdef JVMCI +#if INCLUDE_JVMCI int _pending_deoptimization; oop _pending_failed_speculation; bool _pending_monitorenter; @@ -85,7 +85,7 @@ bool has_pending_exception() const { return _pending_exception != NULL; } const char* exception_file() const { return _exception_file; } int exception_line() const { return _exception_line; } -#ifdef JVMCI +#if INCLUDE_JVMCI int pending_deoptimization() const { return _pending_deoptimization; } oop pending_failed_speculation() const { return _pending_failed_speculation; } bool has_pending_monitorenter() const { return _pending_monitorenter; } @@ -93,7 +93,7 @@ // Code generation support static ByteSize pending_exception_offset() { return byte_offset_of(ThreadShadow, _pending_exception); } -#ifdef JVMCI +#if INCLUDE_JVMCI static ByteSize pending_deoptimization_offset() { return byte_offset_of(ThreadShadow, _pending_deoptimization); } static ByteSize pending_monitorenter_offset() { return byte_offset_of(ThreadShadow, _pending_monitorenter); } static ByteSize pending_failed_speculation_offset() { return byte_offset_of(ThreadShadow, _pending_failed_speculation); } @@ -112,7 +112,7 @@ ThreadShadow() : _pending_exception(NULL), _exception_file(NULL), _exception_line(0) -#ifdef JVMCI +#if INCLUDE_JVMCI , _pending_monitorenter(false), _pending_deoptimization(-1), _pending_failed_speculation(NULL), _pending_transfer_to_interpreter(false) #endif {}
--- a/src/share/vm/utilities/macros.hpp Wed Jul 22 15:50:08 2015 +0200 +++ b/src/share/vm/utilities/macros.hpp Wed Jul 22 08:56:03 2015 -0700 @@ -164,6 +164,20 @@ #define INCLUDE_TRACE 1 #endif // INCLUDE_TRACE +#ifndef INCLUDE_JVMCI +#define INCLUDE_JVMCI 1 +#endif + +#if INCLUDE_JVMCI +#define JVMCI_ONLY(code) code +#define NOT_JVMCI(code) +#define IS_JVMCI_DEFINED true +#else +#define JVMCI_ONLY(code) +#define NOT_JVMCI(code) code +#define IS_JVMCI_DEFINED false +#endif // JVMCI + // COMPILER1 variant #ifdef COMPILER1 #if defined(COMPILER2) || defined(COMPILERJVMCI) @@ -191,20 +205,10 @@ #define NOT_COMPILERJVMCI(code) code #endif // COMPILERJVMCI -#if defined(COMPILERJVMCI) && !defined(JVMCI) -#error "COMPILERJVMCI needs JVMCI to be defined" +#if defined(COMPILERJVMCI) && !INCLUDE_JVMCI +#error "COMPILERJVMCI needs INCLUDE_JVMCI=1" #endif -#ifdef JVMCI -#define JVMCI_ONLY(code) code -#define NOT_JVMCI(code) -#define IS_JVMCI_DEFINED true -#else // JVMCI -#define JVMCI_ONLY(code) -#define NOT_JVMCI(code) code -#define IS_JVMCI_DEFINED false -#endif // JVMCI - #ifdef TIERED #define TIERED_ONLY(code) code #define NOT_TIERED(code)