From 7ed72d943b8d4c5cd0d3707c0c95148db74401bf Mon Sep 17 00:00:00 2001 From: Axel Boldt-Christmas Date: Tue, 23 Sep 2025 07:12:27 +0000 Subject: [PATCH] 8368212: ZGC: Fix spelling and typos in comments Reviewed-by: jsikstro, stefank --- src/hotspot/share/gc/z/zArguments.cpp | 2 +- src/hotspot/share/gc/z/zBarrierSet.cpp | 2 +- src/hotspot/share/gc/z/zDirector.cpp | 2 +- src/hotspot/share/gc/z/zHeapIterator.cpp | 2 +- src/hotspot/share/gc/z/zJNICritical.cpp | 2 +- src/hotspot/share/gc/z/zNMethod.cpp | 2 +- 6 files changed, 6 insertions(+), 6 deletions(-) diff --git a/src/hotspot/share/gc/z/zArguments.cpp b/src/hotspot/share/gc/z/zArguments.cpp index 03aa8e5acee..2435c6fb31f 100644 --- a/src/hotspot/share/gc/z/zArguments.cpp +++ b/src/hotspot/share/gc/z/zArguments.cpp @@ -45,7 +45,7 @@ void ZArguments::initialize_heap_flags_and_sizes() { !FLAG_IS_CMDLINE(SoftMaxHeapSize)) { // We are really just guessing how much memory the program needs. // When that is the case, we don't want the soft and hard limits to be the same - // as it can cause flakyness in the number of GC threads used, in order to keep + // as it can cause flakiness in the number of GC threads used, in order to keep // to a random number we just pulled out of thin air. FLAG_SET_ERGO(SoftMaxHeapSize, MaxHeapSize * 90 / 100); } diff --git a/src/hotspot/share/gc/z/zBarrierSet.cpp b/src/hotspot/share/gc/z/zBarrierSet.cpp index c71c404712c..31996def9c1 100644 --- a/src/hotspot/share/gc/z/zBarrierSet.cpp +++ b/src/hotspot/share/gc/z/zBarrierSet.cpp @@ -117,7 +117,7 @@ static void deoptimize_allocation(JavaThread* thread) { const nmethod* const nm = caller_frame.cb()->as_nmethod(); if ((nm->is_compiled_by_c2() || nm->is_compiled_by_jvmci()) && !caller_frame.is_deoptimized_frame()) { // The JIT might have elided barriers on this object so deoptimize the frame and let the - // intepreter deal with it. + // interpreter deal with it. Deoptimization::deoptimize_frame(thread, caller_frame.id()); } } diff --git a/src/hotspot/share/gc/z/zDirector.cpp b/src/hotspot/share/gc/z/zDirector.cpp index 380a1e48ab4..7c5bb907edf 100644 --- a/src/hotspot/share/gc/z/zDirector.cpp +++ b/src/hotspot/share/gc/z/zDirector.cpp @@ -302,7 +302,7 @@ static bool is_young_small(const ZDirectorStats& stats) { // If the freeable memory isn't even 5% of the heap, we can't expect to free up // all that much memory, so let's not even try - it will likely be a wasted effort - // that takes away CPU power to the hopefullt more profitable major colelction. + // that takes away CPU power to the hopefully more profitable major collection. return young_used_percent <= 5.0; } diff --git a/src/hotspot/share/gc/z/zHeapIterator.cpp b/src/hotspot/share/gc/z/zHeapIterator.cpp index 8e423fe0fca..d6289178ea8 100644 --- a/src/hotspot/share/gc/z/zHeapIterator.cpp +++ b/src/hotspot/share/gc/z/zHeapIterator.cpp @@ -364,7 +364,7 @@ public: virtual void do_nmethod(nmethod* nm) { // If ClassUnloading is turned off, all nmethods are considered strong, // not only those on the call stacks. The heap iteration might happen - // before the concurrent processign of the code cache, make sure that + // before the concurrent processing of the code cache, make sure that // all nmethods have been processed before visiting the oops. _bs_nm->nmethod_entry_barrier(nm); diff --git a/src/hotspot/share/gc/z/zJNICritical.cpp b/src/hotspot/share/gc/z/zJNICritical.cpp index 01c0a0dd6b8..28cc70cb3b5 100644 --- a/src/hotspot/share/gc/z/zJNICritical.cpp +++ b/src/hotspot/share/gc/z/zJNICritical.cpp @@ -160,7 +160,7 @@ void ZJNICritical::exit_inner() { // and we should signal that all Java threads have now exited the // critical region and we are now blocked. if (count == -2) { - // Nofity blocked + // Notify blocked ZLocker locker(_lock); _lock->notify_all(); } diff --git a/src/hotspot/share/gc/z/zNMethod.cpp b/src/hotspot/share/gc/z/zNMethod.cpp index a5d94fc1e5e..780bc9e3bf7 100644 --- a/src/hotspot/share/gc/z/zNMethod.cpp +++ b/src/hotspot/share/gc/z/zNMethod.cpp @@ -314,7 +314,7 @@ oop ZNMethod::oop_load_phantom(const nmethod* nm, int index) { oop ZNMethod::oop_load(const nmethod* const_nm, int index, bool keep_alive) { // The rest of the code is not ready to handle const nmethod, so cast it away - // until we are more consistent with our const corectness. + // until we are more consistent with our const correctness. nmethod* nm = const_cast(const_nm); if (!is_armed(nm)) {