mirror of
https://github.com/openjdk/jdk.git
synced 2026-04-16 01:40:31 +00:00
388 lines
14 KiB
C++
388 lines
14 KiB
C++
/*
|
|
* Copyright (c) 2020, 2026, Oracle and/or its affiliates. All rights reserved.
|
|
* Copyright (c) 2020, 2023, Huawei Technologies Co., Ltd. All rights reserved.
|
|
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
|
|
*
|
|
* This code is free software; you can redistribute it and/or modify it
|
|
* under the terms of the GNU General Public License version 2 only, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This code is distributed in the hope that it will be useful, but WITHOUT
|
|
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
|
|
* version 2 for more details (a copy is included in the LICENSE file that
|
|
* accompanied this code).
|
|
*
|
|
* You should have received a copy of the GNU General Public License version
|
|
* 2 along with this work; if not, write to the Free Software Foundation,
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
|
|
*
|
|
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
|
|
* or visit www.oracle.com if you need additional information or have any
|
|
* questions.
|
|
*
|
|
*/
|
|
|
|
#include "asm/macroAssembler.hpp"
|
|
#include "code/codeBlob.hpp"
|
|
#include "code/codeCache.hpp"
|
|
#include "code/vmreg.inline.hpp"
|
|
#include "compiler/oopMap.hpp"
|
|
#include "logging/logStream.hpp"
|
|
#include "memory/resourceArea.hpp"
|
|
#include "prims/downcallLinker.hpp"
|
|
#include "runtime/globals.hpp"
|
|
#include "runtime/stubCodeGenerator.hpp"
|
|
|
|
#define __ _masm->
|
|
|
|
static const int native_invoker_code_base_size = 256;
|
|
static const int native_invoker_size_per_arg = 8;
|
|
|
|
RuntimeStub* DowncallLinker::make_downcall_stub(BasicType* signature,
|
|
int num_args,
|
|
BasicType ret_bt,
|
|
const ABIDescriptor& abi,
|
|
const GrowableArray<VMStorage>& input_registers,
|
|
const GrowableArray<VMStorage>& output_registers,
|
|
bool needs_return_buffer,
|
|
int captured_state_mask,
|
|
bool needs_transition) {
|
|
int code_size = native_invoker_code_base_size + (num_args * native_invoker_size_per_arg);
|
|
int locs_size = 1; // must be non-zero
|
|
CodeBuffer code("nep_invoker_blob", code_size, locs_size);
|
|
if (code.blob() == nullptr) {
|
|
return nullptr;
|
|
}
|
|
StubGenerator g(&code, signature, num_args, ret_bt, abi,
|
|
input_registers, output_registers,
|
|
needs_return_buffer, captured_state_mask,
|
|
needs_transition);
|
|
g.generate();
|
|
code.log_section_sizes("nep_invoker_blob");
|
|
|
|
bool caller_must_gc_arguments = false;
|
|
bool alloc_fail_is_fatal = false;
|
|
RuntimeStub* stub =
|
|
RuntimeStub::new_runtime_stub("nep_invoker_blob",
|
|
&code,
|
|
g.frame_complete(),
|
|
g.framesize(),
|
|
g.oop_maps(),
|
|
caller_must_gc_arguments,
|
|
alloc_fail_is_fatal);
|
|
if (stub == nullptr) {
|
|
return nullptr;
|
|
}
|
|
|
|
#ifndef PRODUCT
|
|
LogTarget(Trace, foreign, downcall) lt;
|
|
if (lt.is_enabled()) {
|
|
LogStream ls(lt);
|
|
stub->print_on(&ls);
|
|
}
|
|
#endif
|
|
|
|
return stub;
|
|
}
|
|
|
|
static constexpr int FP_BIAS = 0; // sender_sp_offset is 0 on RISCV
|
|
|
|
void DowncallLinker::StubGenerator::pd_add_offset_to_oop(VMStorage reg_oop, VMStorage reg_offset, VMStorage tmp1, VMStorage tmp2) const {
|
|
Register r_tmp1 = as_Register(tmp1);
|
|
Register r_tmp2 = as_Register(tmp2);
|
|
if (reg_oop.is_reg()) {
|
|
assert(reg_oop.type() == StorageType::INTEGER, "expected");
|
|
Register reg_oop_reg = as_Register(reg_oop);
|
|
if (reg_offset.is_reg()) {
|
|
assert(reg_offset.type() == StorageType::INTEGER, "expected");
|
|
__ add(reg_oop_reg, reg_oop_reg, as_Register(reg_offset));
|
|
} else {
|
|
assert(reg_offset.is_stack(), "expected");
|
|
assert(reg_offset.stack_size() == 8, "expected long");
|
|
__ ld(r_tmp1, Address(fp, FP_BIAS + reg_offset.offset()));
|
|
__ add(reg_oop_reg, reg_oop_reg, r_tmp1);
|
|
}
|
|
} else {
|
|
assert(reg_oop.is_stack(), "expected");
|
|
assert(reg_oop.stack_size() == 8, "expected long");
|
|
assert(reg_offset.is_stack(), "expected");
|
|
assert(reg_offset.stack_size() == 8, "expected long");
|
|
__ ld(r_tmp1, Address(fp, FP_BIAS + reg_offset.offset()));
|
|
__ ld(r_tmp2, Address(fp, FP_BIAS + reg_oop.offset()));
|
|
__ add(r_tmp1, r_tmp2, r_tmp1);
|
|
__ sd(r_tmp1, Address(fp, FP_BIAS + reg_oop.offset()));
|
|
}
|
|
}
|
|
|
|
void DowncallLinker::StubGenerator::generate() {
|
|
enum layout {
|
|
fp_off,
|
|
fp_off2,
|
|
ra_off,
|
|
ra_off2,
|
|
framesize // inclusive of return address
|
|
// The following are also computed dynamically:
|
|
// spill area for return value
|
|
// out arg area (e.g. for stack args)
|
|
};
|
|
|
|
GrowableArray<VMStorage> java_regs;
|
|
ForeignGlobals::java_calling_convention(_signature, _num_args, java_regs);
|
|
bool has_objects = false;
|
|
GrowableArray<VMStorage> filtered_java_regs = ForeignGlobals::downcall_filter_offset_regs(java_regs, _signature,
|
|
_num_args, has_objects);
|
|
assert(!(_needs_transition && has_objects), "can not pass objects when doing transition");
|
|
|
|
int allocated_frame_size = 0;
|
|
assert(_abi._shadow_space_bytes == 0, "not expecting shadow space on RISCV64");
|
|
allocated_frame_size += ForeignGlobals::compute_out_arg_bytes(_input_registers);
|
|
|
|
bool should_save_return_value = !_needs_return_buffer;
|
|
RegSpiller out_reg_spiller(_output_registers);
|
|
int out_spill_offset = -1;
|
|
|
|
if (should_save_return_value) {
|
|
out_spill_offset = 0;
|
|
// spill area can be shared with shadow space and out args,
|
|
// since they are only used before the call,
|
|
// and spill area is only used after.
|
|
allocated_frame_size = out_reg_spiller.spill_size_bytes() > allocated_frame_size
|
|
? out_reg_spiller.spill_size_bytes()
|
|
: allocated_frame_size;
|
|
}
|
|
|
|
StubLocations locs;
|
|
locs.set(StubLocations::TARGET_ADDRESS, _abi._scratch1);
|
|
if (_needs_return_buffer) {
|
|
locs.set_frame_data(StubLocations::RETURN_BUFFER, allocated_frame_size);
|
|
allocated_frame_size += BytesPerWord; // for address spill
|
|
}
|
|
if (_captured_state_mask != 0) {
|
|
locs.set_frame_data(StubLocations::CAPTURED_STATE_BUFFER, allocated_frame_size);
|
|
allocated_frame_size += BytesPerWord;
|
|
}
|
|
|
|
// The space we have allocated will look like:
|
|
//
|
|
// FP-> | |
|
|
// |---------------------| = frame_bottom_offset = frame_size
|
|
// | (optional) |
|
|
// | in_reg_spiller area |
|
|
// |---------------------|
|
|
// | (optional) |
|
|
// | capture state buf |
|
|
// |---------------------| = StubLocations::CAPTURED_STATE_BUFFER
|
|
// | (optional) |
|
|
// | return buffer |
|
|
// |---------------------| = StubLocations::RETURN_BUFFER
|
|
// SP-> | out/stack args | or | out_reg_spiller area |
|
|
//
|
|
// Note how the last chunk can be shared, since the 3 uses occur at different times.
|
|
|
|
VMStorage shuffle_reg = as_VMStorage(x9);
|
|
GrowableArray<VMStorage> out_regs = ForeignGlobals::replace_place_holders(_input_registers, locs);
|
|
ArgumentShuffle arg_shuffle(filtered_java_regs, out_regs, shuffle_reg);
|
|
|
|
// Need to spill for state capturing runtime call.
|
|
// The area spilled into is distinct from the capture state buffer.
|
|
RegSpiller in_reg_spiller(out_regs);
|
|
int in_spill_offset = -1;
|
|
if (_captured_state_mask != 0) {
|
|
// The spill area cannot be shared with the out_spill since
|
|
// spilling needs to happen before the call. Allocate a new
|
|
// region in the stack for this spill space.
|
|
in_spill_offset = allocated_frame_size;
|
|
allocated_frame_size += in_reg_spiller.spill_size_bytes();
|
|
}
|
|
|
|
#ifndef PRODUCT
|
|
LogTarget(Trace, foreign, downcall) lt;
|
|
if (lt.is_enabled()) {
|
|
LogStream ls(lt);
|
|
arg_shuffle.print_on(&ls);
|
|
}
|
|
#endif
|
|
|
|
allocated_frame_size = align_up(allocated_frame_size, 16);
|
|
// _frame_size_slots is in 32-bit stack slots:
|
|
_frame_size_slots += framesize + (allocated_frame_size >> LogBytesPerInt);
|
|
assert(is_even(_frame_size_slots / 2), "sp not 16-byte aligned");
|
|
|
|
_oop_maps = _needs_transition ? new OopMapSet() : nullptr;
|
|
address start = __ pc();
|
|
|
|
__ enter();
|
|
|
|
// ra and fp are already in place
|
|
__ sub(sp, sp, allocated_frame_size); // prolog
|
|
|
|
_frame_complete = __ pc() - start; // frame build complete.
|
|
|
|
if (_needs_transition) {
|
|
__ block_comment("{ thread java2native");
|
|
address the_pc = __ pc();
|
|
__ set_last_Java_frame(sp, fp, the_pc, t0);
|
|
OopMap* map = new OopMap(_frame_size_slots, 0);
|
|
_oop_maps->add_gc_map(the_pc - start, map);
|
|
|
|
// State transition
|
|
__ mv(t0, _thread_in_native);
|
|
__ membar(MacroAssembler::LoadStore | MacroAssembler::StoreStore);
|
|
__ sw(t0, Address(xthread, JavaThread::thread_state_offset()));
|
|
__ block_comment("} thread java2native");
|
|
}
|
|
|
|
if (has_objects) {
|
|
add_offsets_to_oops(java_regs, as_VMStorage(t0), as_VMStorage(t1));
|
|
}
|
|
|
|
__ block_comment("{ argument shuffle");
|
|
arg_shuffle.generate(_masm, shuffle_reg, 0, _abi._shadow_space_bytes);
|
|
__ block_comment("} argument shuffle");
|
|
|
|
if (_captured_state_mask != 0) {
|
|
assert(in_spill_offset != -1, "must be");
|
|
__ block_comment("{ load initial thread local");
|
|
in_reg_spiller.generate_spill(_masm, in_spill_offset);
|
|
|
|
// Copy the contents of the capture state buffer into thread local
|
|
__ ld(c_rarg0, Address(sp, locs.data_offset(StubLocations::CAPTURED_STATE_BUFFER)));
|
|
__ mv(c_rarg1, _captured_state_mask);
|
|
__ rt_call(CAST_FROM_FN_PTR(address, DowncallLinker::capture_state_pre));
|
|
|
|
in_reg_spiller.generate_fill(_masm, in_spill_offset);
|
|
__ block_comment("} load initial thread local");
|
|
}
|
|
|
|
__ jalr(as_Register(locs.get(StubLocations::TARGET_ADDRESS)));
|
|
// this call is assumed not to have killed xthread
|
|
|
|
if (_needs_return_buffer) {
|
|
// when use return buffer, copy content of return registers to return buffer,
|
|
// then operations created in BoxBindingCalculator will be operated.
|
|
__ ld(t0, Address(sp, locs.data_offset(StubLocations::RETURN_BUFFER)));
|
|
int offset = 0;
|
|
for (int i = 0; i < _output_registers.length(); i++) {
|
|
VMStorage reg = _output_registers.at(i);
|
|
if (reg.type() == StorageType::INTEGER) {
|
|
__ sd(as_Register(reg), Address(t0, offset));
|
|
offset += 8;
|
|
} else if (reg.type() == StorageType::FLOAT) {
|
|
__ fsd(as_FloatRegister(reg), Address(t0, offset));
|
|
offset += 8;
|
|
} else {
|
|
ShouldNotReachHere();
|
|
}
|
|
}
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////////////
|
|
|
|
if (_captured_state_mask != 0) {
|
|
__ block_comment("{ save thread local");
|
|
|
|
if (should_save_return_value) {
|
|
out_reg_spiller.generate_spill(_masm, out_spill_offset);
|
|
}
|
|
|
|
__ ld(c_rarg0, Address(sp, locs.data_offset(StubLocations::CAPTURED_STATE_BUFFER)));
|
|
__ mv(c_rarg1, _captured_state_mask);
|
|
__ rt_call(CAST_FROM_FN_PTR(address, DowncallLinker::capture_state_post));
|
|
|
|
if (should_save_return_value) {
|
|
out_reg_spiller.generate_fill(_masm, out_spill_offset);
|
|
}
|
|
|
|
__ block_comment("} save thread local");
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////////////
|
|
|
|
Label L_after_safepoint_poll;
|
|
Label L_safepoint_poll_slow_path;
|
|
Label L_reguard;
|
|
Label L_after_reguard;
|
|
if (_needs_transition) {
|
|
// Restore cpu control state after JNI call
|
|
__ restore_cpu_control_state_after_jni(t0);
|
|
|
|
__ block_comment("{ thread native2java");
|
|
__ mv(t0, _thread_in_native_trans);
|
|
__ sw(t0, Address(xthread, JavaThread::thread_state_offset()));
|
|
|
|
// Force this write out before the read below
|
|
if (!UseSystemMemoryBarrier) {
|
|
__ membar(MacroAssembler::AnyAny);
|
|
}
|
|
|
|
__ safepoint_poll(L_safepoint_poll_slow_path, true /* at_return */, false /* in_nmethod */);
|
|
__ lwu(t0, Address(xthread, JavaThread::suspend_flags_offset()));
|
|
__ bnez(t0, L_safepoint_poll_slow_path);
|
|
|
|
__ bind(L_after_safepoint_poll);
|
|
|
|
// change thread state
|
|
__ mv(t0, _thread_in_Java);
|
|
__ membar(MacroAssembler::LoadStore | MacroAssembler::StoreStore);
|
|
__ sw(t0, Address(xthread, JavaThread::thread_state_offset()));
|
|
|
|
__ block_comment("reguard stack check");
|
|
__ lbu(t0, Address(xthread, JavaThread::stack_guard_state_offset()));
|
|
__ mv(t1, StackOverflow::stack_guard_yellow_reserved_disabled);
|
|
__ beq(t0, t1, L_reguard);
|
|
__ bind(L_after_reguard);
|
|
|
|
__ reset_last_Java_frame(true);
|
|
__ block_comment("} thread native2java");
|
|
}
|
|
|
|
__ leave(); // required for proper stackwalking of RuntimeStub frame
|
|
__ ret();
|
|
|
|
//////////////////////////////////////////////////////////////////////////////
|
|
|
|
if (_needs_transition) {
|
|
__ block_comment("{ L_safepoint_poll_slow_path");
|
|
__ bind(L_safepoint_poll_slow_path);
|
|
|
|
if (should_save_return_value) {
|
|
// Need to save the native result registers around any runtime calls.
|
|
out_reg_spiller.generate_spill(_masm, out_spill_offset);
|
|
}
|
|
|
|
__ mv(c_rarg0, xthread);
|
|
assert(frame::arg_reg_save_area_bytes == 0, "not expecting frame reg save area");
|
|
__ rt_call(CAST_FROM_FN_PTR(address, JavaThread::check_special_condition_for_native_trans));
|
|
|
|
if (should_save_return_value) {
|
|
out_reg_spiller.generate_fill(_masm, out_spill_offset);
|
|
}
|
|
__ j(L_after_safepoint_poll);
|
|
__ block_comment("} L_safepoint_poll_slow_path");
|
|
|
|
//////////////////////////////////////////////////////////////////////////////
|
|
|
|
__ block_comment("{ L_reguard");
|
|
__ bind(L_reguard);
|
|
|
|
if (should_save_return_value) {
|
|
// Need to save the native result registers around any runtime calls.
|
|
out_reg_spiller.generate_spill(_masm, out_spill_offset);
|
|
}
|
|
|
|
__ rt_call(CAST_FROM_FN_PTR(address, SharedRuntime::reguard_yellow_pages));
|
|
|
|
if (should_save_return_value) {
|
|
out_reg_spiller.generate_fill(_masm, out_spill_offset);
|
|
}
|
|
|
|
__ j(L_after_reguard);
|
|
__ block_comment("} L_reguard");
|
|
}
|
|
|
|
//////////////////////////////////////////////////////////////////////////////
|
|
|
|
__ flush();
|
|
}
|