@@ -574,8 +574,13 @@ void C2_MacroAssembler::fast_lock(Register objReg, Register boxReg, Register tmp
574
574
// Unconditionally set box->_displaced_header = markWord::unused_mark().
575
575
// Without cast to int32_t this style of movptr will destroy r10 which is typically obj.
576
576
movptr (Address (boxReg, 0 ), (int32_t )intptr_t (markWord::unused_mark ().value ()));
577
- // Intentional fall-through into DONE_LABEL ...
578
577
// Propagate ICC.ZF from CAS above into DONE_LABEL.
578
+ jcc (Assembler::equal, DONE_LABEL); // CAS above succeeded; propagate ZF = 1 (success)
579
+
580
+ cmpptr (r15_thread, rax); // Check if we are already the owner (recursive lock)
581
+ jcc (Assembler::notEqual, DONE_LABEL); // If not recursive, ZF = 0 at this point (fail)
582
+ incq (Address (scrReg, OM_OFFSET_NO_MONITOR_VALUE_TAG (recursions)));
583
+ xorq (rax, rax); // Set ZF = 1 (success) for recursive lock, denoting locking success
579
584
#endif // _LP64
580
585
#if INCLUDE_RTM_OPT
581
586
} // use_rtm()
@@ -670,10 +675,6 @@ void C2_MacroAssembler::fast_unlock(Register objReg, Register boxReg, Register t
670
675
// Refer to the comments in synchronizer.cpp for how we might encode extra
671
676
// state in _succ so we can avoid fetching EntryList|cxq.
672
677
//
673
- // I'd like to add more cases in fast_lock() and fast_unlock() --
674
- // such as recursive enter and exit -- but we have to be wary of
675
- // I$ bloat, T$ effects and BP$ effects.
676
- //
677
678
// If there's no contention try a 1-0 exit. That is, exit without
678
679
// a costly MEMBAR or CAS. See synchronizer.cpp for details on how
679
680
// we detect and recover from the race that the 1-0 exit admits.
@@ -721,9 +722,16 @@ void C2_MacroAssembler::fast_unlock(Register objReg, Register boxReg, Register t
721
722
bind (CheckSucc);
722
723
#else // _LP64
723
724
// It's inflated
724
- xorptr (boxReg, boxReg);
725
- orptr (boxReg, Address (tmpReg, OM_OFFSET_NO_MONITOR_VALUE_TAG (recursions)));
726
- jccb (Assembler::notZero, DONE_LABEL);
725
+ Label LNotRecursive, LSuccess, LGoSlowPath;
726
+
727
+ cmpptr (Address (tmpReg, OM_OFFSET_NO_MONITOR_VALUE_TAG (recursions)), 0 );
728
+ jccb (Assembler::equal, LNotRecursive);
729
+
730
+ // Recursive inflated unlock
731
+ decq (Address (tmpReg, OM_OFFSET_NO_MONITOR_VALUE_TAG (recursions)));
732
+ jmpb (LSuccess);
733
+
734
+ bind (LNotRecursive);
727
735
movptr (boxReg, Address (tmpReg, OM_OFFSET_NO_MONITOR_VALUE_TAG (cxq)));
728
736
orptr (boxReg, Address (tmpReg, OM_OFFSET_NO_MONITOR_VALUE_TAG (EntryList)));
729
737
jccb (Assembler::notZero, CheckSucc);
@@ -732,7 +740,6 @@ void C2_MacroAssembler::fast_unlock(Register objReg, Register boxReg, Register t
732
740
jmpb (DONE_LABEL);
733
741
734
742
// Try to avoid passing control into the slow_path ...
735
- Label LSuccess, LGoSlowPath ;
736
743
bind (CheckSucc);
737
744
738
745
// The following optional optimization can be elided if necessary
0 commit comments