--- a/src/hotspot/cpu/aarch64/assembler_aarch64.cpp	2021-06-08 13:25:01.000000000 +0100
+++ b/src/hotspot/cpu/aarch64/assembler_aarch64.cpp	2026-04-20 22:39:43.585139555 +0100
@@ -280,7 +280,7 @@
 
 bool Assembler::operand_valid_for_add_sub_immediate(int64_t imm) {
   bool shift = false;
-  uint64_t uimm = (uint64_t)uabs((jlong)imm);
+  uint64_t uimm = (uint64_t)g_uabs((jlong)imm);
   if (uimm < (1 << 12))
     return true;
   if (uimm < (1 << 24)
--- a/src/hotspot/cpu/aarch64/assembler_aarch64.hpp	2021-06-08 13:25:01.000000000 +0100
+++ b/src/hotspot/cpu/aarch64/assembler_aarch64.hpp	2026-04-20 22:39:19.541271187 +0100
@@ -889,7 +889,7 @@
   static const uint64_t branch_range = NOT_DEBUG(128 * M) DEBUG_ONLY(2 * M);
 
   static bool reachable_from_branch_at(address branch, address target) {
-    return uabs(target - branch) < branch_range;
+    return g_uabs(target - branch) < branch_range;
   }
 
   // Unconditional branch (immediate)
--- a/src/hotspot/cpu/aarch64/macroAssembler_aarch64.cpp	2021-06-08 13:25:01.000000000 +0100
+++ b/src/hotspot/cpu/aarch64/macroAssembler_aarch64.cpp	2026-04-20 22:39:19.543925018 +0100
@@ -2282,7 +2282,7 @@
   if (operand_valid_for_add_sub_immediate((int)imm)) {
     (this->*insn1)(Rd, Rn, imm);
   } else {
-    if (uabs(imm) < (1 << 24)) {
+    if (g_uabs(imm) < (1 << 24)) {
        (this->*insn1)(Rd, Rn, imm & -(1 << 12));
        (this->*insn1)(Rd, Rd, imm & ((1 << 12)-1));
     } else {
--- a/src/hotspot/cpu/aarch64/stubGenerator_aarch64.cpp	2021-06-08 13:25:01.000000000 +0100
+++ b/src/hotspot/cpu/aarch64/stubGenerator_aarch64.cpp	2026-04-20 22:40:08.219367353 +0100
@@ -1037,7 +1037,7 @@
 
   void copy_memory_small(Register s, Register d, Register count, Register tmp, int step) {
     bool is_backwards = step < 0;
-    size_t granularity = uabs(step);
+    size_t granularity = g_uabs(step);
     int direction = is_backwards ? -1 : 1;
     int unit = wordSize * direction;
 
@@ -1093,7 +1093,7 @@
                    Register count, Register tmp, int step) {
     copy_direction direction = step < 0 ? copy_backwards : copy_forwards;
     bool is_backwards = step < 0;
-    unsigned int granularity = uabs(step);
+    unsigned int granularity = g_uabs(step);
     const Register t0 = r3, t1 = r4;
 
     // <= 80 (or 96 for SIMD) bytes do inline. Direction doesn't matter because we always
--- a/src/hotspot/share/opto/mulnode.cpp	2021-06-08 13:25:01.000000000 +0100
+++ b/src/hotspot/share/opto/mulnode.cpp	2026-04-20 22:39:19.548492071 +0100
@@ -192,7 +192,7 @@
   // Check for negative constant; if so negate the final result
   bool sign_flip = false;
 
-  unsigned int abs_con = uabs(con);
+  unsigned int abs_con = g_uabs(con);
   if (abs_con != (unsigned int)con) {
     sign_flip = true;
   }
@@ -288,7 +288,7 @@
 
   // Check for negative constant; if so negate the final result
   bool sign_flip = false;
-  julong abs_con = uabs(con);
+  julong abs_con = g_uabs(con);
   if (abs_con != (julong)con) {
     sign_flip = true;
   }
--- a/src/hotspot/share/utilities/globalDefinitions.hpp	2021-06-08 13:25:01.000000000 +0100
+++ b/src/hotspot/share/utilities/globalDefinitions.hpp	2026-04-20 22:39:19.549445971 +0100
@@ -1100,7 +1100,7 @@
 
 // abs methods which cannot overflow and so are well-defined across
 // the entire domain of integer types.
-static inline unsigned int uabs(unsigned int n) {
+static inline unsigned int g_uabs(unsigned int n) {
   union {
     unsigned int result;
     int value;
@@ -1109,7 +1109,7 @@
   if (value < 0) result = 0-result;
   return result;
 }
-static inline julong uabs(julong n) {
+static inline julong g_uabs(julong n) {
   union {
     julong result;
     jlong value;
@@ -1118,8 +1118,8 @@
   if (value < 0) result = 0-result;
   return result;
 }
-static inline julong uabs(jlong n) { return uabs((julong)n); }
-static inline unsigned int uabs(int n) { return uabs((unsigned int)n); }
+static inline julong g_uabs(jlong n) { return g_uabs((julong)n); }
+static inline unsigned int g_uabs(int n) { return g_uabs((unsigned int)n); }
 
 // "to" should be greater than "from."
 inline intx byte_size(void* from, void* to) {
