Compare commits

...

9 Commits

Author SHA1 Message Date
Xiaohong Gong
b0ad331809 8371603: C2: Missing Ideal optimizations for load and store vectors on SVE
Reviewed-by: epeter, erfang
Backport-of: b6732d6048259de68a3dd5b4f66ac82f87270404
2025-12-11 02:53:01 +00:00
Kevin Walls
7dc8f786fe 8373111: Test java/lang/management/MemoryMXBean/MemoryManagement.java timed out
Reviewed-by: shade
Backport-of: 9c91c68d1d5938d7e2b9a90c82b0a36ef1a063cd
2025-12-10 19:10:26 +00:00
Xiaolong Peng
15b5789f55 8373116: Genshen: arraycopy_work should be always done for arrays in old gen during young concurrent marking
8372498: [genshen] gc/TestAllocHumongousFragment.java#generational causes intermittent SIGSEGV crashes

Reviewed-by: shade, wkemper
Backport-of: c9ab330b7bdd3cc2410ffdb336a63aa0ac7256a3
2025-12-10 17:34:07 +00:00
Erik Joelsson
1de6f4f2b6 8373255: Unexpected iobj and ipdb files after JDK-8370438
Reviewed-by: serb
Backport-of: 831fe94c75c407b2399be9b89630d8d117c2996c
2025-12-10 14:01:43 +00:00
Emanuel Peter
42fc4fe7b1 8367028: compiler/c2/irTests/TestFloat16ScalarOperations.java failing intermittently because of constant folding
Reviewed-by: thartmann, mhaessig
Backport-of: a4eb57c5ec6254e59e486042015dd00457284ef2
2025-12-10 11:34:41 +00:00
SendaoYan
ec6beaa20a 8373086: Make isexceeded001.java more robust
Reviewed-by: ayang, jsikstro
Backport-of: 4d696d0d0ed523e3c99c68214586673913b1c7b5
2025-12-10 02:15:09 +00:00
Tobias Hartmann
bf0bc37924 8373145: [BACKOUT] Remove ThreadLocalAllocBuffer::_reserve_for_allocation_prefetch
Reviewed-by: chagedorn
Backport-of: 43787890291d71de61b28b8a4e3bf9aaba46757a
2025-12-09 18:37:25 +00:00
Albert Mingkun Yang
22fe70770a 8370198: Test gc/arguments/TestShrinkHeapInSteps.java crashed: assert(left >= right) failed: avoid underflow
Reviewed-by: stefank
Backport-of: d34ef196c298aa91f8511714cfb04b15ae7fbf0a
2025-12-09 18:24:44 +00:00
David Holmes
de686f8a38 8373099: Problem list intermittently failing test sun/awt/image/bug8038000.java
Reviewed-by: kvn, prr
Backport-of: 8e653d394e45180e16714124ed6584f912eb5cba
2025-12-04 22:01:10 +00:00
27 changed files with 746 additions and 86 deletions

View File

@ -234,6 +234,9 @@ define SetupLinkerFlags
ifeq ($(call isTargetOs, macosx), true)
$1_EXTRA_LDFLAGS += -Wl,-object_path_lto,$$($1_OBJECT_DIR)/$$($1_NAME)_lto_helper.o
endif
ifeq ($(TOOLCHAIN_TYPE), microsoft)
$1_EXTRA_LDFLAGS += -LTCGOUT:$$($1_OBJECT_DIR)/$$($1_NAME).iobj
endif
endif
$1_EXTRA_LDFLAGS += $$($1_LDFLAGS_$(OPENJDK_TARGET_OS_TYPE)) $$($1_LDFLAGS_$(OPENJDK_TARGET_OS)) \

View File

@ -346,8 +346,14 @@ source %{
}
bool Matcher::vector_needs_partial_operations(Node* node, const TypeVect* vt) {
// Only SVE has partial vector operations
if (UseSVE == 0) {
// 1. Only SVE requires partial vector operations.
// 2. The vector size in bytes must be smaller than MaxVectorSize.
// 3. Predicated vectors have a mask input, which guarantees that
// out-of-bounds lanes remain inactive.
int length_in_bytes = vt->length_in_bytes();
if (UseSVE == 0 ||
length_in_bytes == MaxVectorSize ||
node->is_predicated_vector()) {
return false;
}
@ -370,21 +376,22 @@ source %{
return !node->in(1)->is_Con();
case Op_LoadVector:
case Op_StoreVector:
// We use NEON load/store instructions if the vector length is <= 128 bits.
return vt->length_in_bytes() > 16;
case Op_AddReductionVI:
case Op_AddReductionVL:
// We may prefer using NEON instructions rather than SVE partial operations.
return !VM_Version::use_neon_for_vector(vt->length_in_bytes());
// For these ops, we prefer using NEON instructions rather than SVE
// predicated instructions for better performance.
return !VM_Version::use_neon_for_vector(length_in_bytes);
case Op_MinReductionV:
case Op_MaxReductionV:
// For BYTE/SHORT/INT/FLOAT/DOUBLE types, we may prefer using NEON
// instructions rather than SVE partial operations.
// For BYTE/SHORT/INT/FLOAT/DOUBLE types, we prefer using NEON
// instructions rather than SVE predicated instructions for
// better performance.
return vt->element_basic_type() == T_LONG ||
!VM_Version::use_neon_for_vector(vt->length_in_bytes());
!VM_Version::use_neon_for_vector(length_in_bytes);
default:
// For other ops whose vector size is smaller than the max vector size, a
// full-sized unpredicated operation does not impact the final vector result.
// For other ops whose vector size is smaller than the max vector
// size, a full-sized unpredicated operation does not impact the
// vector result.
return false;
}
}

View File

@ -336,8 +336,14 @@ source %{
}
bool Matcher::vector_needs_partial_operations(Node* node, const TypeVect* vt) {
// Only SVE has partial vector operations
if (UseSVE == 0) {
// 1. Only SVE requires partial vector operations.
// 2. The vector size in bytes must be smaller than MaxVectorSize.
// 3. Predicated vectors have a mask input, which guarantees that
// out-of-bounds lanes remain inactive.
int length_in_bytes = vt->length_in_bytes();
if (UseSVE == 0 ||
length_in_bytes == MaxVectorSize ||
node->is_predicated_vector()) {
return false;
}
@ -360,21 +366,22 @@ source %{
return !node->in(1)->is_Con();
case Op_LoadVector:
case Op_StoreVector:
// We use NEON load/store instructions if the vector length is <= 128 bits.
return vt->length_in_bytes() > 16;
case Op_AddReductionVI:
case Op_AddReductionVL:
// We may prefer using NEON instructions rather than SVE partial operations.
return !VM_Version::use_neon_for_vector(vt->length_in_bytes());
// For these ops, we prefer using NEON instructions rather than SVE
// predicated instructions for better performance.
return !VM_Version::use_neon_for_vector(length_in_bytes);
case Op_MinReductionV:
case Op_MaxReductionV:
// For BYTE/SHORT/INT/FLOAT/DOUBLE types, we may prefer using NEON
// instructions rather than SVE partial operations.
// For BYTE/SHORT/INT/FLOAT/DOUBLE types, we prefer using NEON
// instructions rather than SVE predicated instructions for
// better performance.
return vt->element_basic_type() == T_LONG ||
!VM_Version::use_neon_for_vector(vt->length_in_bytes());
!VM_Version::use_neon_for_vector(length_in_bytes);
default:
// For other ops whose vector size is smaller than the max vector size, a
// full-sized unpredicated operation does not impact the final vector result.
// For other ops whose vector size is smaller than the max vector
// size, a full-sized unpredicated operation does not impact the
// vector result.
return false;
}
}

View File

@ -5379,7 +5379,6 @@ void MacroAssembler::set_narrow_klass(Register dst, Klass* k) {
assert (UseCompressedClassPointers, "should only be used for compressed headers");
assert (oop_recorder() != nullptr, "this assembler needs an OopRecorder");
int index = oop_recorder()->find_index(k);
assert(! Universe::heap()->is_in(k), "should not be an oop");
InstructionMark im(this);
RelocationHolder rspec = metadata_Relocation::spec(index);

View File

@ -6335,8 +6335,36 @@ instruct loadConD_Ex(regD dst, immD src) %{
// Prefetch instructions.
// Must be safe to execute with invalid address (cannot fault).
// Special prefetch versions which use the dcbz instruction.
instruct prefetch_alloc_zero(indirectMemory mem, iRegLsrc src) %{
match(PrefetchAllocation (AddP mem src));
predicate(AllocatePrefetchStyle == 3);
ins_cost(MEMORY_REF_COST);
format %{ "PREFETCH $mem, 2, $src \t// Prefetch write-many with zero" %}
size(4);
ins_encode %{
__ dcbz($src$$Register, $mem$$base$$Register);
%}
ins_pipe(pipe_class_memory);
%}
instruct prefetch_alloc_zero_no_offset(indirectMemory mem) %{
match(PrefetchAllocation mem);
predicate(AllocatePrefetchStyle == 3);
ins_cost(MEMORY_REF_COST);
format %{ "PREFETCH $mem, 2 \t// Prefetch write-many with zero" %}
size(4);
ins_encode %{
__ dcbz($mem$$base$$Register);
%}
ins_pipe(pipe_class_memory);
%}
instruct prefetch_alloc(indirectMemory mem, iRegLsrc src) %{
match(PrefetchAllocation (AddP mem src));
predicate(AllocatePrefetchStyle != 3);
ins_cost(MEMORY_REF_COST);
format %{ "PREFETCH $mem, 2, $src \t// Prefetch write-many" %}
@ -6349,6 +6377,7 @@ instruct prefetch_alloc(indirectMemory mem, iRegLsrc src) %{
instruct prefetch_alloc_no_offset(indirectMemory mem) %{
match(PrefetchAllocation mem);
predicate(AllocatePrefetchStyle != 3);
ins_cost(MEMORY_REF_COST);
format %{ "PREFETCH $mem, 2 \t// Prefetch write-many" %}

View File

@ -4933,7 +4933,6 @@ void MacroAssembler::set_narrow_klass(Register dst, Klass* k) {
assert (UseCompressedClassPointers, "should only be used for compressed headers");
assert (oop_recorder() != nullptr, "this assembler needs an OopRecorder");
int index = oop_recorder()->find_index(k);
assert(!Universe::heap()->is_in(k), "should not be an oop");
narrowKlass nk = CompressedKlassPointers::encode(k);
relocate(metadata_Relocation::spec(index), [&] {

View File

@ -630,6 +630,14 @@ bool SerialHeap::requires_barriers(stackChunkOop obj) const {
// Returns "TRUE" iff "p" points into the committed areas of the heap.
bool SerialHeap::is_in(const void* p) const {
// precondition
verify_not_in_native_if_java_thread();
if (!is_in_reserved(p)) {
// If it's not even in reserved.
return false;
}
return _young_gen->is_in(p) || _old_gen->is_in(p);
}
@ -797,3 +805,12 @@ void SerialHeap::gc_epilogue(bool full) {
MetaspaceCounters::update_performance_counters();
};
#ifdef ASSERT
void SerialHeap::verify_not_in_native_if_java_thread() {
if (Thread::current()->is_Java_thread()) {
JavaThread* thread = JavaThread::current();
assert(thread->thread_state() != _thread_in_native, "precondition");
}
}
#endif

View File

@ -111,6 +111,8 @@ private:
void print_tracing_info() const override;
void stop() override {};
static void verify_not_in_native_if_java_thread() NOT_DEBUG_RETURN;
public:
// Returns JNI_OK on success
jint initialize() override;

View File

@ -37,6 +37,7 @@
#include "utilities/copy.hpp"
size_t ThreadLocalAllocBuffer::_max_size = 0;
int ThreadLocalAllocBuffer::_reserve_for_allocation_prefetch = 0;
unsigned int ThreadLocalAllocBuffer::_target_refills = 0;
ThreadLocalAllocBuffer::ThreadLocalAllocBuffer() :
@ -224,6 +225,30 @@ void ThreadLocalAllocBuffer::startup_initialization() {
// abort during VM initialization.
_target_refills = MAX2(_target_refills, 2U);
#ifdef COMPILER2
// If the C2 compiler is present, extra space is needed at the end of
// TLABs, otherwise prefetching instructions generated by the C2
// compiler will fault (due to accessing memory outside of heap).
// The amount of space is the max of the number of lines to
// prefetch for array and for instance allocations. (Extra space must be
// reserved to accommodate both types of allocations.)
//
// Only SPARC-specific BIS instructions are known to fault. (Those
// instructions are generated if AllocatePrefetchStyle==3 and
// AllocatePrefetchInstr==1). To be on the safe side, however,
// extra space is reserved for all combinations of
// AllocatePrefetchStyle and AllocatePrefetchInstr.
//
// If the C2 compiler is not present, no space is reserved.
// +1 for rounding up to next cache line, +1 to be safe
if (CompilerConfig::is_c2_or_jvmci_compiler_enabled()) {
int lines = MAX2(AllocatePrefetchLines, AllocateInstancePrefetchLines) + 2;
_reserve_for_allocation_prefetch = (AllocatePrefetchDistance + AllocatePrefetchStepSize * lines) /
(int)HeapWordSize;
}
#endif
// During jvm startup, the main thread is initialized
// before the heap is initialized. So reinitialize it now.
guarantee(Thread::current()->is_Java_thread(), "tlab initialization thread not Java thread");
@ -429,7 +454,8 @@ void ThreadLocalAllocStats::publish() {
}
size_t ThreadLocalAllocBuffer::end_reserve() {
return CollectedHeap::lab_alignment_reserve();
size_t reserve_size = CollectedHeap::lab_alignment_reserve();
return MAX2(reserve_size, (size_t)_reserve_for_allocation_prefetch);
}
const HeapWord* ThreadLocalAllocBuffer::start_relaxed() const {

View File

@ -58,6 +58,7 @@ private:
size_t _allocated_before_last_gc; // total bytes allocated up until the last gc
static size_t _max_size; // maximum size of any TLAB
static int _reserve_for_allocation_prefetch; // Reserve at the end of the TLAB
static unsigned _target_refills; // expected number of refills between GCs
unsigned _number_of_refills;

View File

@ -128,8 +128,8 @@ public:
void write_ref_array(HeapWord* start, size_t count);
private:
template <class T>
inline void arraycopy_marking(T* dst, size_t count);
template <bool IS_GENERATIONAL, class T>
void arraycopy_marking(T* dst, size_t count);
template <class T>
inline void arraycopy_evacuation(T* src, size_t count);
template <class T>

View File

@ -429,7 +429,11 @@ void ShenandoahBarrierSet::arraycopy_barrier(T* src, T* dst, size_t count) {
// If marking old or young, we must evaluate the SATB barrier. This will be the only
// action if we are not marking old. If we are marking old, we must still evaluate the
// load reference barrier for a young collection.
arraycopy_marking(dst, count);
if (_heap->mode()->is_generational()) {
arraycopy_marking<true>(dst, count);
} else {
arraycopy_marking<false>(dst, count);
}
}
if ((gc_state & ShenandoahHeap::EVACUATION) != 0) {
@ -441,11 +445,12 @@ void ShenandoahBarrierSet::arraycopy_barrier(T* src, T* dst, size_t count) {
}
}
template <class T>
template <bool IS_GENERATIONAL, class T>
void ShenandoahBarrierSet::arraycopy_marking(T* dst, size_t count) {
assert(_heap->is_concurrent_mark_in_progress(), "only during marking");
if (ShenandoahSATBBarrier) {
if (!_heap->marking_context()->allocated_after_mark_start(reinterpret_cast<HeapWord*>(dst))) {
if (!_heap->marking_context()->allocated_after_mark_start(reinterpret_cast<HeapWord*>(dst)) ||
(IS_GENERATIONAL && _heap->heap_region_containing(dst)->is_old() && _heap->is_concurrent_young_mark_in_progress())) {
arraycopy_work<T, false, false, true>(dst, count);
}
}

View File

@ -1914,7 +1914,8 @@ Node* PhaseMacroExpand::prefetch_allocation(Node* i_o, Node*& needgc_false,
transform_later(cache_adr);
cache_adr = new CastP2XNode(needgc_false, cache_adr);
transform_later(cache_adr);
// Address is aligned to execute prefetch to the beginning of cache line size.
// Address is aligned to execute prefetch to the beginning of cache line size
// (it is important when BIS instruction is used on SPARC as prefetch).
Node* mask = _igvn.MakeConX(~(intptr_t)(step_size-1));
cache_adr = new AndXNode(cache_adr, mask);
transform_later(cache_adr);

View File

@ -329,6 +329,10 @@ public:
static bool match_rule_supported_vector_masked(int opcode, int vlen, BasicType bt);
// Determines if a vector operation needs to be partially implemented with a mask
// controlling only the lanes in range [0, vector_length) are processed. This applies
// to operations whose vector length is less than the hardware-supported maximum
// vector length. Returns true if the operation requires masking, false otherwise.
static bool vector_needs_partial_operations(Node* node, const TypeVect* vt);
static bool vector_rearrange_requires_load_shuffle(BasicType elem_bt, int vlen);

View File

@ -936,28 +936,26 @@ bool VectorNode::is_scalar_op_that_returns_int_but_vector_op_returns_long(int op
}
}
// Idealize vector operations whose vector size is less than the hardware supported
// max vector size. Generate a vector mask for the operation. Lanes with indices
// inside of the vector size are set to true, while the remaining lanes are set to
// false. Returns the corresponding masked vector node.
static Node* ideal_partial_operations(PhaseGVN* phase, Node* node, const TypeVect* vt) {
if (!Matcher::vector_needs_partial_operations(node, vt)) {
return nullptr;
}
Node* VectorNode::try_to_gen_masked_vector(PhaseGVN* gvn, Node* node, const TypeVect* vt) {
int vopc = node->Opcode();
uint vlen = vt->length();
BasicType bt = vt->element_basic_type();
assert(Matcher::match_rule_supported_vector_masked(vopc, vlen, bt),
"The masked feature is required for the vector operation");
assert(Matcher::match_rule_supported_vector(Op_VectorMaskGen, vlen, bt),
"'VectorMaskGen' is required to generate a vector mask");
// Predicated vectors do not need to add another mask input
if (node->is_predicated_vector() || !Matcher::has_predicated_vectors() ||
!Matcher::match_rule_supported_vector_masked(vopc, vlen, bt) ||
!Matcher::match_rule_supported_vector(Op_VectorMaskGen, vlen, bt)) {
return nullptr;
}
Node* mask = nullptr;
// Generate a vector mask for vector operation whose vector length is lower than the
// hardware supported max vector length.
if (vt->length_in_bytes() < (uint)MaxVectorSize) {
Node* length = gvn->transform(new ConvI2LNode(gvn->makecon(TypeInt::make(vlen))));
mask = gvn->transform(VectorMaskGenNode::make(length, bt, vlen));
} else {
return nullptr;
}
// Generate a vector mask, with lanes inside of the vector length set to true.
Node* length = phase->transform(new ConvI2LNode(phase->makecon(TypeInt::make(vlen))));
Node* mask = phase->transform(VectorMaskGenNode::make(length, bt, vlen));
// Generate the related masked op for vector load/store/load_gather/store_scatter.
// Or append the mask to the vector op's input list by default.
@ -1037,8 +1035,9 @@ bool VectorNode::should_swap_inputs_to_help_global_value_numbering() {
}
Node* VectorNode::Ideal(PhaseGVN* phase, bool can_reshape) {
if (Matcher::vector_needs_partial_operations(this, vect_type())) {
return try_to_gen_masked_vector(phase, this, vect_type());
Node* n = ideal_partial_operations(phase, this, vect_type());
if (n != nullptr) {
return n;
}
// Sort inputs of commutative non-predicated vector operations to help value numbering.
@ -1119,9 +1118,9 @@ LoadVectorNode* LoadVectorNode::make(int opc, Node* ctl, Node* mem,
}
Node* LoadVectorNode::Ideal(PhaseGVN* phase, bool can_reshape) {
const TypeVect* vt = vect_type();
if (Matcher::vector_needs_partial_operations(this, vt)) {
return VectorNode::try_to_gen_masked_vector(phase, this, vt);
Node* n = ideal_partial_operations(phase, this, vect_type());
if (n != nullptr) {
return n;
}
return LoadNode::Ideal(phase, can_reshape);
}
@ -1133,9 +1132,9 @@ StoreVectorNode* StoreVectorNode::make(int opc, Node* ctl, Node* mem, Node* adr,
}
Node* StoreVectorNode::Ideal(PhaseGVN* phase, bool can_reshape) {
const TypeVect* vt = vect_type();
if (Matcher::vector_needs_partial_operations(this, vt)) {
return VectorNode::try_to_gen_masked_vector(phase, this, vt);
Node* n = ideal_partial_operations(phase, this, vect_type());
if (n != nullptr) {
return n;
}
return StoreNode::Ideal(phase, can_reshape);
}
@ -1411,11 +1410,11 @@ ReductionNode* ReductionNode::make(int opc, Node* ctrl, Node* n1, Node* n2, Basi
}
Node* ReductionNode::Ideal(PhaseGVN* phase, bool can_reshape) {
const TypeVect* vt = vect_type();
if (Matcher::vector_needs_partial_operations(this, vt)) {
return VectorNode::try_to_gen_masked_vector(phase, this, vt);
Node* n = ideal_partial_operations(phase, this, vect_type());
if (n != nullptr) {
return n;
}
return nullptr;
return Node::Ideal(phase, can_reshape);
}
// Convert fromLong to maskAll if the input sets or unsets all lanes.
@ -1893,11 +1892,11 @@ Node* VectorMaskOpNode::make(Node* mask, const Type* ty, int mopc) {
}
Node* VectorMaskOpNode::Ideal(PhaseGVN* phase, bool can_reshape) {
const TypeVect* vt = vect_type();
if (Matcher::vector_needs_partial_operations(this, vt)) {
return VectorNode::try_to_gen_masked_vector(phase, this, vt);
Node* n = ideal_partial_operations(phase, this, vect_type());
if (n != nullptr) {
return n;
}
return nullptr;
return TypeNode::Ideal(phase, can_reshape);
}
Node* VectorMaskCastNode::Identity(PhaseGVN* phase) {

View File

@ -117,7 +117,6 @@ class VectorNode : public TypeNode {
static bool is_vector_bitwise_not_pattern(Node* n);
static Node* degenerate_vector_rotate(Node* n1, Node* n2, bool is_rotate_left, int vlen,
BasicType bt, PhaseGVN* phase);
static Node* try_to_gen_masked_vector(PhaseGVN* gvn, Node* node, const TypeVect* vt);
// [Start, end) half-open range defining which operands are vectors
static void vector_operands(Node* n, uint* start, uint* end);

View File

@ -353,6 +353,7 @@
nonstatic_field(ThreadLocalAllocBuffer, _pf_top, HeapWord*) \
nonstatic_field(ThreadLocalAllocBuffer, _desired_size, size_t) \
nonstatic_field(ThreadLocalAllocBuffer, _refill_waste_limit, size_t) \
static_field(ThreadLocalAllocBuffer, _reserve_for_allocation_prefetch, int) \
static_field(ThreadLocalAllocBuffer, _target_refills, unsigned) \
nonstatic_field(ThreadLocalAllocBuffer, _number_of_refills, unsigned) \
nonstatic_field(ThreadLocalAllocBuffer, _refill_waste, unsigned) \

View File

@ -76,9 +76,10 @@ public class ThreadLocalAllocBuffer extends VMObject {
private long endReserve() {
long labAlignmentReserve = VM.getVM().getLabAlignmentReserve();
long reserveForAllocationPrefetch = VM.getVM().getReserveForAllocationPrefetch();
long heapWordSize = VM.getVM().getHeapWordSize();
return labAlignmentReserve * heapWordSize;
return Math.max(labAlignmentReserve, reserveForAllocationPrefetch) * heapWordSize;
}
/** Support for iteration over heap -- not sure how this will

View File

@ -123,6 +123,7 @@ public class VM {
private int invocationEntryBCI;
private ReversePtrs revPtrs;
private VMRegImpl vmregImpl;
private int reserveForAllocationPrefetch;
private int labAlignmentReserve;
// System.getProperties from debuggee VM
@ -446,6 +447,8 @@ public class VM {
boolType = (CIntegerType) db.lookupType("bool");
Type threadLocalAllocBuffer = db.lookupType("ThreadLocalAllocBuffer");
CIntegerField reserveForAllocationPrefetchField = threadLocalAllocBuffer.getCIntegerField("_reserve_for_allocation_prefetch");
reserveForAllocationPrefetch = (int)reserveForAllocationPrefetchField.getCInteger(intType);
Type collectedHeap = db.lookupType("CollectedHeap");
CIntegerField labAlignmentReserveField = collectedHeap.getCIntegerField("_lab_alignment_reserve");
@ -912,6 +915,10 @@ public class VM {
return vmInternalInfo;
}
public int getReserveForAllocationPrefetch() {
return reserveForAllocationPrefetch;
}
public int getLabAlignmentReserve() {
return labAlignmentReserve;
}

View File

@ -187,9 +187,3 @@ vmTestbase/nsk/jdwp/ThreadReference/ForceEarlyReturn/forceEarlyReturn001/forceEa
vmTestbase/nsk/monitoring/ThreadMXBean/ThreadInfo/Multi/Multi005/TestDescription.java 8076494 windows-x64
vmTestbase/nsk/monitoring/ThreadMXBean/findMonitorDeadlockedThreads/find006/TestDescription.java 8310144 macosx-aarch64
vmTestbase/nsk/monitoring/MemoryPoolMBean/isUsageThresholdExceeded/isexceeded001/TestDescription.java 8373022 generic-all
vmTestbase/nsk/monitoring/MemoryPoolMBean/isUsageThresholdExceeded/isexceeded002/TestDescription.java 8373022 generic-all
vmTestbase/nsk/monitoring/MemoryPoolMBean/isUsageThresholdExceeded/isexceeded003/TestDescription.java 8373022 generic-all
vmTestbase/nsk/monitoring/MemoryPoolMBean/isUsageThresholdExceeded/isexceeded004/TestDescription.java 8373022 generic-all
vmTestbase/nsk/monitoring/MemoryPoolMBean/isUsageThresholdExceeded/isexceeded005/TestDescription.java 8373022 generic-all

View File

@ -68,6 +68,15 @@ public class TestFloat16ScalarOperations {
private static final Float16 RANDOM4 = Float16.valueOf(genF.next());
private static final Float16 RANDOM5 = Float16.valueOf(genF.next());
// We have to ensure that the constants are not special values that lead the operations to
// constant fold. For example "x + 0" could constant fold to "x", so we need to avoid that
// the add constant is zero.
private static Generator<Float> genSmallRangeF = G.uniformFloats(0.1f, 0.9f);
private static final Float16 RANDOM_CON_ADD = Float16.valueOf(genSmallRangeF.next());
private static final Float16 RANDOM_CON_SUB = Float16.valueOf(genSmallRangeF.next());
private static final Float16 RANDOM_CON_MUL = Float16.valueOf(genSmallRangeF.next());
private static final Float16 RANDOM_CON_DIV = Float16.valueOf(genSmallRangeF.next());
private static Float16 RANDOM1_VAR = RANDOM1;
private static Float16 RANDOM2_VAR = RANDOM2;
private static Float16 RANDOM3_VAR = RANDOM3;
@ -435,10 +444,10 @@ public class TestFloat16ScalarOperations {
@Warmup(10000)
public short testRandomFP16ConstantPatternSet1() {
short res = 0;
res += Float.floatToFloat16(RANDOM1_VAR.floatValue() + RANDOM2.floatValue());
res += Float.floatToFloat16(RANDOM2_VAR.floatValue() - RANDOM3.floatValue());
res += Float.floatToFloat16(RANDOM3_VAR.floatValue() * RANDOM4.floatValue());
res += Float.floatToFloat16(RANDOM4_VAR.floatValue() / RANDOM5.floatValue());
res += Float.floatToFloat16(RANDOM1_VAR.floatValue() + RANDOM_CON_ADD.floatValue());
res += Float.floatToFloat16(RANDOM2_VAR.floatValue() - RANDOM_CON_SUB.floatValue());
res += Float.floatToFloat16(RANDOM3_VAR.floatValue() * RANDOM_CON_MUL.floatValue());
res += Float.floatToFloat16(RANDOM4_VAR.floatValue() / RANDOM_CON_DIV.floatValue());
return res;
}
@ -456,10 +465,10 @@ public class TestFloat16ScalarOperations {
@Warmup(10000)
public short testRandomFP16ConstantPatternSet2() {
short res = 0;
res += Float.floatToFloat16(RANDOM2.floatValue() + RANDOM1_VAR.floatValue());
res += Float.floatToFloat16(RANDOM3.floatValue() - RANDOM2_VAR.floatValue());
res += Float.floatToFloat16(RANDOM4.floatValue() * RANDOM3_VAR.floatValue());
res += Float.floatToFloat16(RANDOM5.floatValue() / RANDOM4_VAR.floatValue());
res += Float.floatToFloat16(RANDOM_CON_ADD.floatValue() + RANDOM1_VAR.floatValue());
res += Float.floatToFloat16(RANDOM_CON_SUB.floatValue() - RANDOM2_VAR.floatValue());
res += Float.floatToFloat16(RANDOM_CON_MUL.floatValue() * RANDOM3_VAR.floatValue());
res += Float.floatToFloat16(RANDOM_CON_DIV.floatValue() / RANDOM4_VAR.floatValue());
return res;
}

View File

@ -1448,6 +1448,16 @@ public class IRNode {
beforeMatchingNameRegex(VECTOR_MASK_LANE_IS_SET, "ExtractUB");
}
public static final String VECTOR_MASK_GEN = PREFIX + "VECTOR_MASK_GEN" + POSTFIX;
static {
beforeMatchingNameRegex(VECTOR_MASK_GEN, "VectorMaskGen");
}
public static final String VECTOR_MASK_FIRST_TRUE = PREFIX + "VECTOR_MASK_FIRST_TRUE" + POSTFIX;
static {
beforeMatchingNameRegex(VECTOR_MASK_FIRST_TRUE, "VectorMaskFirstTrue");
}
// Can only be used if avx512_vnni is available.
public static final String MUL_ADD_VS2VI_VNNI = PREFIX + "MUL_ADD_VS2VI_VNNI" + POSTFIX;
static {

View File

@ -0,0 +1,105 @@
/*
* Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*/
package compiler.vectorapi;
import compiler.lib.generators.*;
import compiler.lib.ir_framework.*;
import jdk.incubator.vector.*;
import jdk.test.lib.Asserts;
/**
* @test 8371603
* @key randomness
* @library /test/lib /
* @summary Test the missing optimization issues for vector load/store caused by JDK-8286941
* @modules jdk.incubator.vector
*
* @run driver ${test.main.class}
*/
public class TestVectorLoadStoreOptimization {
private static final int LENGTH = 1024;
private static final Generators random = Generators.G;
private static final VectorSpecies<Integer> SPECIES = IntVector.SPECIES_PREFERRED;
private static int[] a;
static {
a = new int[LENGTH];
random.fill(random.ints(), a);
}
// Test that "LoadVectorNode::Ideal()" calls "LoadNode::Ideal()" as expected,
// which sees the previous stores that go to the same position in-dependently,
// and optimize out the load with matched store values.
@Test
@IR(counts = { IRNode.LOAD_VECTOR_I, "1" },
applyIfCPUFeatureOr = {"asimd", "true", "avx", "true", "rvv", "true"})
public static void testLoadVector() {
IntVector v1 = IntVector.fromArray(SPECIES, a, 0);
v1.intoArray(a, SPECIES.length());
v1.intoArray(a, 2 * SPECIES.length());
// The second load vector equals to the first one and should be optimized
// out by "LoadNode::Ideal()".
IntVector v2 = IntVector.fromArray(SPECIES, a, SPECIES.length());
v2.intoArray(a, 3 * SPECIES.length());
}
@Check(test = "testLoadVector")
public static void testLoadVectorVerify() {
for (int i = SPECIES.length(); i < 4 * SPECIES.length(); i += SPECIES.length()) {
for (int j = 0; j < SPECIES.length(); j++) {
Asserts.assertEquals(a[i + j], a[j]);
}
}
}
// Test that "StoreVectorNode::Ideal()" calls "StoreNode::Ideal()" as expected,
// which can get rid of previous stores that go to the same position.
@Test
@IR(counts = { IRNode.STORE_VECTOR, "1" },
applyIfCPUFeatureOr = {"asimd", "true", "avx", "true", "rvv", "true"})
public static void testStoreVector() {
IntVector v1 = IntVector.fromArray(SPECIES, a, 0 * SPECIES.length());
IntVector v2 = IntVector.fromArray(SPECIES, a, 1 * SPECIES.length());
// Useless store to same position as below, which should be optimized out by
// "StoreNode::Ideal()".
v1.intoArray(a, 3 * SPECIES.length());
v2.intoArray(a, 3 * SPECIES.length());
}
@Check(test = "testStoreVector")
public static void testStoreVectorVerify() {
for (int i = 3 * SPECIES.length(); i < 4 * SPECIES.length(); i++) {
Asserts.assertEquals(a[i], a[i - 2 * SPECIES.length()]);
}
}
public static void main(String[] args) {
TestFramework testFramework = new TestFramework();
testFramework.setDefaultWarmup(10000)
.addFlags("--add-modules=jdk.incubator.vector")
.start();
}
}

View File

@ -0,0 +1,432 @@
/*
* Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*/
package compiler.vectorapi;
import compiler.lib.generators.*;
import compiler.lib.ir_framework.*;
import jdk.incubator.vector.*;
import jdk.test.lib.Asserts;
/**
* @test 8371603
* @key randomness
* @library /test/lib /
* @summary Test vector operations with vector size less than MaxVectorSize
* @modules jdk.incubator.vector
*
* @run driver ${test.main.class}
*/
public class TestVectorOperationsWithPartialSize {
private static final int SIZE = 1024;
private static final Generators random = Generators.G;
private static final VectorSpecies<Integer> ISPEC_128 = IntVector.SPECIES_128;
private static final VectorSpecies<Long> LSPEC_128 = LongVector.SPECIES_128;
private static final VectorSpecies<Float> FSPEC_128 = FloatVector.SPECIES_128;
private static final VectorSpecies<Double> DSPEC_128 = DoubleVector.SPECIES_128;
private static final VectorSpecies<Integer> ISPEC_256 = IntVector.SPECIES_256;
private static final VectorSpecies<Long> LSPEC_256 = LongVector.SPECIES_256;
private static int[] ia;
private static int[] ib;
private static long[] la;
private static long[] lb;
private static float[] fa;
private static float[] fb;
private static double[] da;
private static double[] db;
private static boolean[] m;
private static boolean[] mr;
private static int[] indices;
static {
ia = new int[SIZE];
ib = new int[SIZE];
la = new long[SIZE];
lb = new long[SIZE];
fa = new float[SIZE];
fb = new float[SIZE];
da = new double[SIZE];
db = new double[SIZE];
m = new boolean[SIZE];
mr = new boolean[SIZE];
indices = new int[SIZE];
random.fill(random.ints(), ia);
random.fill(random.longs(), la);
random.fill(random.floats(), fa);
random.fill(random.doubles(), da);
random.fill(random.uniformInts(0, ISPEC_128.length()), indices);
for (int i = 0; i < SIZE; i++) {
m[i] = i % 2 == 0;
}
}
// ================ Load/Store/Gather/Scatter Tests ==================
private static void verifyLoadStore(int[] expected, int[] actual, int vlen) {
for (int i = 0; i < vlen; i++) {
Asserts.assertEquals(expected[i], actual[i]);
}
}
private static void verifyLoadGatherStoreScatter(int[] expected, int[] actual, int[] indices, int vlen) {
for (int i = 0; i < vlen; i++) {
Asserts.assertEquals(expected[indices[i]], actual[indices[i]]);
}
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "0",
IRNode.LOAD_VECTOR_I, IRNode.VECTOR_SIZE_4, "1",
IRNode.STORE_VECTOR, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public void testLoadStore_128() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
v.intoArray(ib, 0);
verifyLoadStore(ia, ib, ISPEC_128.length());
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.LOAD_VECTOR_MASKED, "1",
IRNode.STORE_VECTOR_MASKED, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=64"})
public void testLoadStore_256() {
IntVector v = IntVector.fromArray(ISPEC_256, ia, 0);
v.intoArray(ib, 0);
verifyLoadStore(ia, ib, ISPEC_256.length());
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.LOAD_VECTOR_GATHER_MASKED, "1",
IRNode.STORE_VECTOR_SCATTER_MASKED, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public void testLoadGatherStoreScatter_128() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0, indices, 0);
v.intoArray(ib, 0, indices, 0);
verifyLoadGatherStoreScatter(ia, ib, indices, ISPEC_128.length());
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.LOAD_VECTOR_GATHER_MASKED, "1",
IRNode.STORE_VECTOR_SCATTER_MASKED, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=64"})
public void testLoadGatherStoreScatter_256() {
IntVector v = IntVector.fromArray(ISPEC_256, ia, 0, indices, 0);
v.intoArray(ib, 0, indices, 0);
verifyLoadGatherStoreScatter(ia, ib, indices, ISPEC_256.length());
}
// ===================== Reduction Tests - Add =====================
interface binOpInt {
int apply(int a, int b);
}
interface binOpLong {
long apply(long a, long b);
}
private static int reduceLanes(int init, int[] arr, int vlen, binOpInt f) {
int result = init;
for (int i = 0; i < vlen; i++) {
result = f.apply(arr[i], result);
}
return result;
}
private static long reduceLanes(long init, long[] arr, int vlen,binOpLong f) {
long result = init;
for (int i = 0; i < vlen; i++) {
result = f.apply(arr[i], result);
}
return result;
}
// Reduction add operations with integer types are implemented with NEON SIMD instructions
// when the vector size is less than or equal to 128-bit.
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "0",
IRNode.ADD_REDUCTION_VI, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public int testAddReductionInt_128() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
int result = v.reduceLanes(VectorOperators.ADD);
Asserts.assertEquals(reduceLanes(0, ia, ISPEC_128.length(), (a, b) -> (a + b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.ADD_REDUCTION_VI, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=64"})
public int testAddReductionInt_256() {
IntVector v = IntVector.fromArray(ISPEC_256, ia, 0);
int result = v.reduceLanes(VectorOperators.ADD);
Asserts.assertEquals(reduceLanes(0, ia, ISPEC_256.length(), (a, b) -> (a + b)), result);
return result;
}
// Reduction add operations with long types are implemented with NEON SIMD instructions
// when the vector size is less than or equal to 128-bit.
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "0",
IRNode.ADD_REDUCTION_VL, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public long testAddReductionLong_128() {
LongVector v = LongVector.fromArray(LSPEC_128, la, 0);
long result = v.reduceLanes(VectorOperators.ADD);
Asserts.assertEquals(reduceLanes(0L, la, LSPEC_128.length(), (a, b) -> (a + b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.ADD_REDUCTION_VL, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=64"})
public long testAddReductionLong_256() {
LongVector v = LongVector.fromArray(LSPEC_256, la, 0);
long result = v.reduceLanes(VectorOperators.ADD);
Asserts.assertEquals(reduceLanes(0L, la, LSPEC_256.length(), (a, b) -> (a + b)), result);
return result;
}
private static void verifyAddReductionFloat(float actual, float[] arr, int vlen) {
float expected = 0.0f;
for (int i = 0; i < vlen; i++) {
expected += arr[i];
}
// Floating point addition reduction ops may introduce rounding errors.
float ROUNDING_ERROR_FACTOR_ADD = 10.0f;
float tolerance = Math.ulp(expected) * ROUNDING_ERROR_FACTOR_ADD;
if (Math.abs(expected - actual) > tolerance) {
throw new RuntimeException(
"assertEqualsWithTolerance" +
": expected " + expected + " but was " + actual +
" (tolerance: " + tolerance + ", diff: " + Math.abs(expected - actual) + ")"
);
}
}
private static void verifyAddReductionDouble(double actual, double[] arr, int vlen) {
double expected = 0.0;
for (int i = 0; i < vlen; i++) {
expected += arr[i];
}
// Floating point addition reduction ops may introduce rounding errors.
double ROUNDING_ERROR_FACTOR_ADD = 10.0;
double tolerance = Math.ulp(expected) * ROUNDING_ERROR_FACTOR_ADD;
if (Math.abs(expected - actual) > tolerance) {
throw new RuntimeException(
"assertEqualsWithTolerance" +
": expected " + expected + " but was " + actual +
" (tolerance: " + tolerance + ", diff: " + Math.abs(expected - actual) + ")"
);
}
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.ADD_REDUCTION_VF, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public float testAddReductionFloat() {
FloatVector v = FloatVector.fromArray(FSPEC_128, fa, 0);
float result = v.reduceLanes(VectorOperators.ADD);
verifyAddReductionFloat(result, fa, FSPEC_128.length());
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.ADD_REDUCTION_VD, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public double testAddReductionDouble() {
DoubleVector v = DoubleVector.fromArray(DSPEC_128, da, 0);
double result = v.reduceLanes(VectorOperators.ADD);
verifyAddReductionDouble(result, da, DSPEC_128.length());
return result;
}
// ============== Reduction Tests - Logical ==============
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.AND_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public int testAndReduction() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
int result = v.reduceLanes(VectorOperators.AND);
Asserts.assertEquals(reduceLanes(-1, ia, ISPEC_128.length(), (a, b) -> (a & b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.OR_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public int testOrReduction() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
int result = v.reduceLanes(VectorOperators.OR);
Asserts.assertEquals(reduceLanes(0, ia, ISPEC_128.length(), (a, b) -> (a | b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.XOR_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">=32"})
public int testXorReduction() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
int result = v.reduceLanes(VectorOperators.XOR);
Asserts.assertEquals(reduceLanes(0, ia, ISPEC_128.length(), (a, b) -> (a ^ b)), result);
return result;
}
// ===================== Reduction Tests - Min/Max =====================
// Reduction min operations with non-long types are implemented with NEON SIMD instructions
// when the vector size is less than or equal to 128-bit.
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "0",
IRNode.MIN_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public int testMinReductionInt_128() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
int result = v.reduceLanes(VectorOperators.MIN);
Asserts.assertEquals(reduceLanes(Integer.MAX_VALUE, ia, ISPEC_128.length(), (a, b) -> Math.min(a, b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.MIN_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 64"})
public int testMinReductionInt_256() {
IntVector v = IntVector.fromArray(ISPEC_256, ia, 0);
int result = v.reduceLanes(VectorOperators.MIN);
Asserts.assertEquals(reduceLanes(Integer.MAX_VALUE, ia, ISPEC_256.length(), (a, b) -> Math.min(a, b)), result);
return result;
}
// Reduction max operations with non-long types are implemented with NEON SIMD instructions
// when the vector size is less than or equal to 128-bit.
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "0",
IRNode.MAX_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public int testMaxReductionInt_128() {
IntVector v = IntVector.fromArray(ISPEC_128, ia, 0);
int result = v.reduceLanes(VectorOperators.MAX);
Asserts.assertEquals(reduceLanes(Integer.MIN_VALUE, ia, ISPEC_128.length(), (a, b) -> Math.max(a, b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.MAX_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 64"})
public int testMaxReductionInt_256() {
IntVector v = IntVector.fromArray(ISPEC_256, ia, 0);
int result = v.reduceLanes(VectorOperators.MAX);
Asserts.assertEquals(reduceLanes(Integer.MIN_VALUE, ia, ISPEC_256.length(), (a, b) -> Math.max(a, b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.MIN_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public static long testMinReductionLong() {
LongVector v = LongVector.fromArray(LSPEC_128, la, 0);
long result = v.reduceLanes(VectorOperators.MIN);
Asserts.assertEquals(reduceLanes(Long.MAX_VALUE, la, LSPEC_128.length(), (a, b) -> Math.min(a, b)), result);
return result;
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.MAX_REDUCTION_V, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public static long testMaxReductionLong() {
LongVector v = LongVector.fromArray(LSPEC_128, la, 0);
long result = v.reduceLanes(VectorOperators.MAX);
Asserts.assertEquals(reduceLanes(Long.MIN_VALUE, la, LSPEC_128.length(), (a, b) -> Math.max(a, b)), result);
return result;
}
// ====================== VectorMask Tests ======================
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.VECTOR_LOAD_MASK, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public static void testLoadMask() {
VectorMask<Integer> vm = VectorMask.fromArray(ISPEC_128, m, 0);
vm.not().intoArray(mr, 0);
// Verify that the mask is loaded correctly.
for (int i = 0; i < ISPEC_128.length(); i++) {
Asserts.assertEquals(!m[i], mr[i]);
}
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.VECTOR_MASK_CMP, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public static void testVectorMaskCmp() {
IntVector v1 = IntVector.fromArray(ISPEC_128, ia, 0);
IntVector v2 = IntVector.fromArray(ISPEC_128, ib, 0);
VectorMask<Integer> vm = v1.compare(VectorOperators.LT, v2);
vm.intoArray(mr, 0);
// Verify that the mask is generated correctly.
for (int i = 0; i < ISPEC_128.length(); i++) {
Asserts.assertEquals(ia[i] < ib[i], mr[i]);
}
}
@Test
@IR(counts = {IRNode.VECTOR_MASK_GEN, "1",
IRNode.VECTOR_MASK_FIRST_TRUE, "1"},
applyIfCPUFeature = {"sve", "true"}, applyIf = {"MaxVectorSize", ">= 32"})
public static int testFirstTrue() {
VectorMask<Integer> vm = ISPEC_128.maskAll(false);
int result = vm.firstTrue();
// The result is the vector length if no lane is true.
// This is the default behavior of the firstTrue method.
Asserts.assertEquals(ISPEC_128.length(), result);
return result;
}
public static void main(String[] args) {
TestFramework testFramework = new TestFramework();
testFramework.setDefaultWarmup(10000)
.addFlags("--add-modules=jdk.incubator.vector")
.start();
}
}

View File

@ -92,7 +92,8 @@ public class isexceeded001 {
// but cannot assume this affects the pool we are testing.
b = new byte[INCREMENT];
isExceeded = monitor.isUsageThresholdExceeded(pool);
// Ensure the observation of isExceeded is sticky to match peakUsage.
isExceeded = isExceeded || monitor.isUsageThresholdExceeded(pool);
log.display(" Allocated heap. isExceeded = " + isExceeded);
// Fetch usage information: use peak usage in comparisons below, in case usage went up and then down.

View File

@ -500,6 +500,7 @@ java/awt/GraphicsDevice/DisplayModes/UnknownRefrshRateTest.java 8286436 macosx-a
java/awt/image/multiresolution/MultiresolutionIconTest.java 8291979 linux-x64,windows-all
java/awt/event/SequencedEvent/MultipleContextsFunctionalTest.java 8305061 macosx-x64
sun/java2d/DirectX/OnScreenRenderingResizeTest/OnScreenRenderingResizeTest.java 8301177 linux-x64
sun/awt/image/bug8038000.java 8373065 generic-all
# Several tests which fail on some hidpi systems/macosx12-aarch64 system
java/awt/Window/8159168/SetShapeTest.java 8274106 macosx-aarch64

View File

@ -1,5 +1,5 @@
/*
* Copyright (c) 2003, 2022, Oracle and/or its affiliates. All rights reserved.
* Copyright (c) 2003, 2025, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
@ -49,7 +49,7 @@
*
* @modules jdk.management
*
* @run main/timeout=600 MemoryManagement
* @run main/othervm/timeout=600 -Xmn8m MemoryManagement
*/
import java.lang.management.*;
@ -58,6 +58,10 @@ import javax.management.*;
import javax.management.openmbean.CompositeData;
public class MemoryManagement {
private static final int YOUNG_GEN_SIZE = 8 * 1024 * 1024; // Must match -Xmn set on the @run line
private static final int NUM_CHUNKS = 2;
private static final MemoryMXBean mm = ManagementFactory.getMemoryMXBean();
private static final List pools =
Collections.synchronizedList(ManagementFactory.getMemoryPoolMXBeans());
@ -66,9 +70,6 @@ public class MemoryManagement {
private static volatile MemoryPoolMXBean mpool = null;
private static volatile boolean trace = false;
private static volatile boolean testFailed = false;
private static final int NUM_CHUNKS = 2;
// Must match -Xmn set on the @run line
private static final int YOUNG_GEN_SIZE = 8 * 1024 * 1024;
private static volatile long chunkSize;
private static volatile int listenerInvoked = 0;