| // Copyright 2015 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // +build ignore |
| |
| // Garbage collector: stack barriers |
| // |
| // Stack barriers enable the garbage collector to determine how much |
| // of a gorountine stack has changed between when a stack is scanned |
| // during the concurrent scan phase and when it is re-scanned during |
| // the stop-the-world mark termination phase. Mark termination only |
| // needs to re-scan the changed part, so for deep stacks this can |
| // significantly reduce GC pause time compared to the alternative of |
| // re-scanning whole stacks. The deeper the stacks, the more stack |
| // barriers help. |
| // |
| // When stacks are scanned during the concurrent scan phase, the stack |
| // scan installs stack barriers by selecting stack frames and |
| // overwriting the saved return PCs (or link registers) of these |
| // frames with the PC of a "stack barrier trampoline". Later, when a |
| // selected frame returns, it "returns" to this trampoline instead of |
| // returning to its actual caller. The trampoline records that the |
| // stack has unwound past this frame and jumps to the original return |
| // PC recorded when the stack barrier was installed. Mark termination |
| // re-scans only as far as the first frame that hasn't hit a stack |
| // barrier and then removes and un-hit stack barriers. |
| // |
| // This scheme is very lightweight. No special code is required in the |
| // mutator to record stack unwinding and the trampoline is only a few |
| // assembly instructions. |
| // |
| // Book-keeping |
| // ------------ |
| // |
| // The primary cost of stack barriers is book-keeping: the runtime has |
| // to record the locations of all stack barriers and the original |
| // return PCs in order to return to the correct caller when a stack |
| // barrier is hit and so it can remove un-hit stack barriers. In order |
| // to minimize this cost, the Go runtime places stack barriers in |
| // exponentially-spaced frames, starting 1K past the current frame. |
| // The book-keeping structure hence grows logarithmically with the |
| // size of the stack and mark termination re-scans at most twice as |
| // much stack as necessary. |
| // |
| // The runtime reserves space for this book-keeping structure at the |
| // top of the stack allocation itself (just above the outermost |
| // frame). This is necessary because the regular memory allocator can |
| // itself grow the stack, and hence can't be used when allocating |
| // stack-related structures. |
| // |
| // For debugging, the runtime also supports installing stack barriers |
| // at every frame. However, this requires significantly more |
| // book-keeping space. |
| // |
| // Correctness |
| // ----------- |
| // |
| // The runtime and the compiler cooperate to ensure that all objects |
| // reachable from the stack as of mark termination are marked. |
| // Anything unchanged since the concurrent scan phase will be marked |
| // because it is marked by the concurrent scan. After the concurrent |
| // scan, there are three possible classes of stack modifications that |
| // must be tracked: |
| // |
| // 1) Mutator writes below the lowest un-hit stack barrier. This |
| // includes all writes performed by an executing function to its own |
| // stack frame. This part of the stack will be re-scanned by mark |
| // termination, which will mark any objects made reachable from |
| // modifications to this part of the stack. |
| // |
| // 2) Mutator writes above the lowest un-hit stack barrier. It's |
| // possible for a mutator to modify the stack above the lowest un-hit |
| // stack barrier if a higher frame has passed down a pointer to a |
| // stack variable in its frame. This is called an "up-pointer". The |
| // compiler ensures that writes through up-pointers have an |
| // accompanying write barrier (it simply doesn't distinguish between |
| // writes through up-pointers and writes through heap pointers). This |
| // write barrier marks any object made reachable from modifications to |
| // this part of the stack. |
| // |
| // 3) Runtime writes to the stack. Various runtime operations such as |
| // sends to unbuffered channels can write to arbitrary parts of the |
| // stack, including above the lowest un-hit stack barrier. We solve |
| // this in two ways. In many cases, the runtime can perform an |
| // explicit write barrier operation like in case 2. However, in the |
| // case of bulk memory move (typedmemmove), the runtime doesn't |
| // necessary have ready access to a pointer bitmap for the memory |
| // being copied, so it simply unwinds any stack barriers below the |
| // destination. |
| // |
| // Gotchas |
| // ------- |
| // |
| // Anything that inspects or manipulates the stack potentially needs |
| // to understand stack barriers. The most obvious case is that |
| // gentraceback needs to use the original return PC when it encounters |
| // the stack barrier trampoline. Anything that unwinds the stack such |
| // as panic/recover must unwind stack barriers in tandem with |
| // unwinding the stack. |
| // |
| // Stack barriers require that any goroutine whose stack has been |
| // scanned must execute write barriers. Go solves this by simply |
| // enabling write barriers globally during the concurrent scan phase. |
| // However, traditionally, write barriers are not enabled during this |
| // phase. |
| // |
| // Synchronization |
| // --------------- |
| // |
| // For the most part, accessing and modifying stack barriers is |
| // synchronized around GC safe points. Installing stack barriers |
| // forces the G to a safe point, while all other operations that |
| // modify stack barriers run on the G and prevent it from reaching a |
| // safe point. |
| // |
| // Subtlety arises when a G may be tracebacked when *not* at a safe |
| // point. This happens during sigprof. For this, each G has a "stack |
| // barrier lock" (see gcLockStackBarriers, gcUnlockStackBarriers). |
| // Operations that manipulate stack barriers acquire this lock, while |
| // sigprof tries to acquire it and simply skips the traceback if it |
| // can't acquire it. There is one exception for performance and |
| // complexity reasons: hitting a stack barrier manipulates the stack |
| // barrier list without acquiring the stack barrier lock. For this, |
| // gentraceback performs a special fix up if the traceback starts in |
| // the stack barrier function. |
| |
| package runtime |
| |
| import ( |
| "runtime/internal/atomic" |
| "runtime/internal/sys" |
| "unsafe" |
| ) |
| |
| const debugStackBarrier = false |
| |
| // firstStackBarrierOffset is the approximate byte offset at |
| // which to place the first stack barrier from the current SP. |
| // This is a lower bound on how much stack will have to be |
| // re-scanned during mark termination. Subsequent barriers are |
| // placed at firstStackBarrierOffset * 2^n offsets. |
| // |
| // For debugging, this can be set to 0, which will install a |
| // stack barrier at every frame. If you do this, you may also |
| // have to raise _StackMin, since the stack barrier |
| // bookkeeping will use a large amount of each stack. |
| var firstStackBarrierOffset = 1024 |
| |
| // gcMaxStackBarriers returns the maximum number of stack barriers |
| // that can be installed in a stack of stackSize bytes. |
| func gcMaxStackBarriers(stackSize int) (n int) { |
| if debug.gcstackbarrieroff > 0 { |
| return 0 |
| } |
| |
| if firstStackBarrierOffset == 0 { |
| // Special debugging case for inserting stack barriers |
| // at every frame. Steal half of the stack for the |
| // []stkbar. Technically, if the stack were to consist |
| // solely of return PCs we would need two thirds of |
| // the stack, but stealing that much breaks things and |
| // this doesn't happen in practice. |
| return stackSize / 2 / int(unsafe.Sizeof(stkbar{})) |
| } |
| |
| offset := firstStackBarrierOffset |
| for offset < stackSize { |
| n++ |
| offset *= 2 |
| } |
| return n + 1 |
| } |
| |
| // gcInstallStackBarrier installs a stack barrier over the return PC of frame. |
| //go:nowritebarrier |
| func gcInstallStackBarrier(gp *g, frame *stkframe) bool { |
| if frame.lr == 0 { |
| if debugStackBarrier { |
| print("not installing stack barrier with no LR, goid=", gp.goid, "\n") |
| } |
| return false |
| } |
| |
| if frame.fn.entry == cgocallback_gofuncPC { |
| // cgocallback_gofunc doesn't return to its LR; |
| // instead, its return path puts LR in g.sched.pc and |
| // switches back to the system stack on which |
| // cgocallback_gofunc was originally called. We can't |
| // have a stack barrier in g.sched.pc, so don't |
| // install one in this frame. |
| if debugStackBarrier { |
| print("not installing stack barrier over LR of cgocallback_gofunc, goid=", gp.goid, "\n") |
| } |
| return false |
| } |
| |
| // Save the return PC and overwrite it with stackBarrier. |
| var lrUintptr uintptr |
| if usesLR { |
| lrUintptr = frame.sp |
| } else { |
| lrUintptr = frame.fp - sys.RegSize |
| } |
| lrPtr := (*sys.Uintreg)(unsafe.Pointer(lrUintptr)) |
| if debugStackBarrier { |
| print("install stack barrier at ", hex(lrUintptr), " over ", hex(*lrPtr), ", goid=", gp.goid, "\n") |
| if uintptr(*lrPtr) != frame.lr { |
| print("frame.lr=", hex(frame.lr)) |
| throw("frame.lr differs from stack LR") |
| } |
| } |
| |
| gp.stkbar = gp.stkbar[:len(gp.stkbar)+1] |
| stkbar := &gp.stkbar[len(gp.stkbar)-1] |
| stkbar.savedLRPtr = lrUintptr |
| stkbar.savedLRVal = uintptr(*lrPtr) |
| *lrPtr = sys.Uintreg(stackBarrierPC) |
| return true |
| } |
| |
| // gcRemoveStackBarriers removes all stack barriers installed in gp's stack. |
| // |
| // gp's stack barriers must be locked. |
| // |
| //go:nowritebarrier |
| func gcRemoveStackBarriers(gp *g) { |
| if debugStackBarrier && gp.stkbarPos != 0 { |
| print("hit ", gp.stkbarPos, " stack barriers, goid=", gp.goid, "\n") |
| } |
| |
| // Remove stack barriers that we didn't hit. |
| for _, stkbar := range gp.stkbar[gp.stkbarPos:] { |
| gcRemoveStackBarrier(gp, stkbar) |
| } |
| |
| // Clear recorded stack barriers so copystack doesn't try to |
| // adjust them. |
| gp.stkbarPos = 0 |
| gp.stkbar = gp.stkbar[:0] |
| } |
| |
| // gcRemoveStackBarrier removes a single stack barrier. It is the |
| // inverse operation of gcInstallStackBarrier. |
| // |
| // This is nosplit to ensure gp's stack does not move. |
| // |
| //go:nowritebarrier |
| //go:nosplit |
| func gcRemoveStackBarrier(gp *g, stkbar stkbar) { |
| if debugStackBarrier { |
| print("remove stack barrier at ", hex(stkbar.savedLRPtr), " with ", hex(stkbar.savedLRVal), ", goid=", gp.goid, "\n") |
| } |
| lrPtr := (*sys.Uintreg)(unsafe.Pointer(stkbar.savedLRPtr)) |
| if val := *lrPtr; val != sys.Uintreg(stackBarrierPC) { |
| printlock() |
| print("at *", hex(stkbar.savedLRPtr), " expected stack barrier PC ", hex(stackBarrierPC), ", found ", hex(val), ", goid=", gp.goid, "\n") |
| print("gp.stkbar=") |
| gcPrintStkbars(gp, -1) |
| print(", gp.stack=[", hex(gp.stack.lo), ",", hex(gp.stack.hi), ")\n") |
| throw("stack barrier lost") |
| } |
| *lrPtr = sys.Uintreg(stkbar.savedLRVal) |
| } |
| |
| // gcTryRemoveAllStackBarriers tries to remove stack barriers from all |
| // Gs in gps. It is best-effort and efficient. If it can't remove |
| // barriers from a G immediately, it will simply skip it. |
| func gcTryRemoveAllStackBarriers(gps []*g) { |
| for _, gp := range gps { |
| retry: |
| for { |
| switch s := readgstatus(gp); s { |
| default: |
| break retry |
| |
| case _Grunnable, _Gsyscall, _Gwaiting: |
| if !castogscanstatus(gp, s, s|_Gscan) { |
| continue |
| } |
| gcLockStackBarriers(gp) |
| gcRemoveStackBarriers(gp) |
| gcUnlockStackBarriers(gp) |
| restartg(gp) |
| break retry |
| } |
| } |
| } |
| } |
| |
| // gcPrintStkbars prints the stack barriers of gp for debugging. It |
| // places a "@@@" marker at gp.stkbarPos. If marker >= 0, it will also |
| // place a "==>" marker before the marker'th entry. |
| func gcPrintStkbars(gp *g, marker int) { |
| print("[") |
| for i, s := range gp.stkbar { |
| if i > 0 { |
| print(" ") |
| } |
| if i == int(gp.stkbarPos) { |
| print("@@@ ") |
| } |
| if i == marker { |
| print("==> ") |
| } |
| print("*", hex(s.savedLRPtr), "=", hex(s.savedLRVal)) |
| } |
| if int(gp.stkbarPos) == len(gp.stkbar) { |
| print(" @@@") |
| } |
| if marker == len(gp.stkbar) { |
| print(" ==>") |
| } |
| print("]") |
| } |
| |
| // gcUnwindBarriers marks all stack barriers up the frame containing |
| // sp as hit and removes them. This is used during stack unwinding for |
| // panic/recover and by heapBitsBulkBarrier to force stack re-scanning |
| // when its destination is on the stack. |
| // |
| // This is nosplit to ensure gp's stack does not move. |
| // |
| //go:nosplit |
| func gcUnwindBarriers(gp *g, sp uintptr) { |
| gcLockStackBarriers(gp) |
| // On LR machines, if there is a stack barrier on the return |
| // from the frame containing sp, this will mark it as hit even |
| // though it isn't, but it's okay to be conservative. |
| before := gp.stkbarPos |
| for int(gp.stkbarPos) < len(gp.stkbar) && gp.stkbar[gp.stkbarPos].savedLRPtr < sp { |
| gcRemoveStackBarrier(gp, gp.stkbar[gp.stkbarPos]) |
| gp.stkbarPos++ |
| } |
| gcUnlockStackBarriers(gp) |
| if debugStackBarrier && gp.stkbarPos != before { |
| print("skip barriers below ", hex(sp), " in goid=", gp.goid, ": ") |
| // We skipped barriers between the "==>" marker |
| // (before) and the "@@@" marker (gp.stkbarPos). |
| gcPrintStkbars(gp, int(before)) |
| print("\n") |
| } |
| } |
| |
| // nextBarrierPC returns the original return PC of the next stack barrier. |
| // Used by getcallerpc, so it must be nosplit. |
| //go:nosplit |
| func nextBarrierPC() uintptr { |
| gp := getg() |
| return gp.stkbar[gp.stkbarPos].savedLRVal |
| } |
| |
| // setNextBarrierPC sets the return PC of the next stack barrier. |
| // Used by setcallerpc, so it must be nosplit. |
| //go:nosplit |
| func setNextBarrierPC(pc uintptr) { |
| gp := getg() |
| gcLockStackBarriers(gp) |
| gp.stkbar[gp.stkbarPos].savedLRVal = pc |
| gcUnlockStackBarriers(gp) |
| } |
| |
| // gcLockStackBarriers synchronizes with tracebacks of gp's stack |
| // during sigprof for installation or removal of stack barriers. It |
| // blocks until any current sigprof is done tracebacking gp's stack |
| // and then disallows profiling tracebacks of gp's stack. |
| // |
| // This is necessary because a sigprof during barrier installation or |
| // removal could observe inconsistencies between the stkbar array and |
| // the stack itself and crash. |
| // |
| //go:nosplit |
| func gcLockStackBarriers(gp *g) { |
| // Disable preemption so scanstack cannot run while the caller |
| // is manipulating the stack barriers. |
| acquirem() |
| for !atomic.Cas(&gp.stackLock, 0, 1) { |
| osyield() |
| } |
| } |
| |
| //go:nosplit |
| func gcTryLockStackBarriers(gp *g) bool { |
| mp := acquirem() |
| result := atomic.Cas(&gp.stackLock, 0, 1) |
| if !result { |
| releasem(mp) |
| } |
| return result |
| } |
| |
| func gcUnlockStackBarriers(gp *g) { |
| atomic.Store(&gp.stackLock, 0) |
| releasem(getg().m) |
| } |