| // Copyright 2009 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // Central free lists. |
| // |
| // See malloc.go for an overview. |
| // |
| // The MCentral doesn't actually contain the list of free objects; the MSpan does. |
| // Each MCentral is two lists of MSpans: those with free objects (c->nonempty) |
| // and those that are completely allocated (c->empty). |
| |
| package runtime |
| |
| // Central list of free objects of a given size. |
| type mcentral struct { |
| lock mutex |
| sizeclass int32 |
| nonempty mspan // list of spans with a free object |
| empty mspan // list of spans with no free objects (or cached in an mcache) |
| } |
| |
| // Initialize a single central free list. |
| func mCentral_Init(c *mcentral, sizeclass int32) { |
| c.sizeclass = sizeclass |
| mSpanList_Init(&c.nonempty) |
| mSpanList_Init(&c.empty) |
| } |
| |
| // Allocate a span to use in an MCache. |
| func mCentral_CacheSpan(c *mcentral) *mspan { |
| // Perform proportional sweep work. We don't directly reuse |
| // the spans we're sweeping here for this allocation because |
| // these can hold any size class. We'll sweep one more span |
| // below and use that because it will have the right size |
| // class and be hot in our cache. |
| pagesOwed := int64(mheap_.sweepPagesPerByte * float64(memstats.heap_live-memstats.heap_marked)) |
| if pagesOwed-int64(mheap_.pagesSwept) > 1 { |
| // Get the debt down to one page, which we're likely |
| // to take care of below (if we don't, that's fine; |
| // we'll pick up the slack later). |
| for pagesOwed-int64(atomicload64(&mheap_.pagesSwept)) > 1 { |
| if gosweepone() == ^uintptr(0) { |
| mheap_.sweepPagesPerByte = 0 |
| break |
| } |
| } |
| } |
| |
| lock(&c.lock) |
| sg := mheap_.sweepgen |
| retry: |
| var s *mspan |
| for s = c.nonempty.next; s != &c.nonempty; s = s.next { |
| if s.sweepgen == sg-2 && cas(&s.sweepgen, sg-2, sg-1) { |
| mSpanList_Remove(s) |
| mSpanList_InsertBack(&c.empty, s) |
| unlock(&c.lock) |
| mSpan_Sweep(s, true) |
| goto havespan |
| } |
| if s.sweepgen == sg-1 { |
| // the span is being swept by background sweeper, skip |
| continue |
| } |
| // we have a nonempty span that does not require sweeping, allocate from it |
| mSpanList_Remove(s) |
| mSpanList_InsertBack(&c.empty, s) |
| unlock(&c.lock) |
| goto havespan |
| } |
| |
| for s = c.empty.next; s != &c.empty; s = s.next { |
| if s.sweepgen == sg-2 && cas(&s.sweepgen, sg-2, sg-1) { |
| // we have an empty span that requires sweeping, |
| // sweep it and see if we can free some space in it |
| mSpanList_Remove(s) |
| // swept spans are at the end of the list |
| mSpanList_InsertBack(&c.empty, s) |
| unlock(&c.lock) |
| mSpan_Sweep(s, true) |
| if s.freelist.ptr() != nil { |
| goto havespan |
| } |
| lock(&c.lock) |
| // the span is still empty after sweep |
| // it is already in the empty list, so just retry |
| goto retry |
| } |
| if s.sweepgen == sg-1 { |
| // the span is being swept by background sweeper, skip |
| continue |
| } |
| // already swept empty span, |
| // all subsequent ones must also be either swept or in process of sweeping |
| break |
| } |
| unlock(&c.lock) |
| |
| // Replenish central list if empty. |
| s = mCentral_Grow(c) |
| if s == nil { |
| return nil |
| } |
| lock(&c.lock) |
| mSpanList_InsertBack(&c.empty, s) |
| unlock(&c.lock) |
| |
| // At this point s is a non-empty span, queued at the end of the empty list, |
| // c is unlocked. |
| havespan: |
| cap := int32((s.npages << _PageShift) / s.elemsize) |
| n := cap - int32(s.ref) |
| if n == 0 { |
| throw("empty span") |
| } |
| if s.freelist.ptr() == nil { |
| throw("freelist empty") |
| } |
| s.incache = true |
| return s |
| } |
| |
| // Return span from an MCache. |
| func mCentral_UncacheSpan(c *mcentral, s *mspan) { |
| lock(&c.lock) |
| |
| s.incache = false |
| |
| if s.ref == 0 { |
| throw("uncaching full span") |
| } |
| |
| cap := int32((s.npages << _PageShift) / s.elemsize) |
| n := cap - int32(s.ref) |
| if n > 0 { |
| mSpanList_Remove(s) |
| mSpanList_Insert(&c.nonempty, s) |
| } |
| unlock(&c.lock) |
| } |
| |
| // Free n objects from a span s back into the central free list c. |
| // Called during sweep. |
| // Returns true if the span was returned to heap. Sets sweepgen to |
| // the latest generation. |
| // If preserve=true, don't return the span to heap nor relink in MCentral lists; |
| // caller takes care of it. |
| func mCentral_FreeSpan(c *mcentral, s *mspan, n int32, start gclinkptr, end gclinkptr, preserve bool) bool { |
| if s.incache { |
| throw("freespan into cached span") |
| } |
| |
| // Add the objects back to s's free list. |
| wasempty := s.freelist.ptr() == nil |
| end.ptr().next = s.freelist |
| s.freelist = start |
| s.ref -= uint16(n) |
| |
| if preserve { |
| // preserve is set only when called from MCentral_CacheSpan above, |
| // the span must be in the empty list. |
| if s.next == nil { |
| throw("can't preserve unlinked span") |
| } |
| atomicstore(&s.sweepgen, mheap_.sweepgen) |
| return false |
| } |
| |
| lock(&c.lock) |
| |
| // Move to nonempty if necessary. |
| if wasempty { |
| mSpanList_Remove(s) |
| mSpanList_Insert(&c.nonempty, s) |
| } |
| |
| // delay updating sweepgen until here. This is the signal that |
| // the span may be used in an MCache, so it must come after the |
| // linked list operations above (actually, just after the |
| // lock of c above.) |
| atomicstore(&s.sweepgen, mheap_.sweepgen) |
| |
| if s.ref != 0 { |
| unlock(&c.lock) |
| return false |
| } |
| |
| // s is completely freed, return it to the heap. |
| mSpanList_Remove(s) |
| s.needzero = 1 |
| s.freelist = 0 |
| unlock(&c.lock) |
| heapBitsForSpan(s.base()).initSpan(s.layout()) |
| mHeap_Free(&mheap_, s, 0) |
| return true |
| } |
| |
| // Fetch a new span from the heap and carve into objects for the free list. |
| func mCentral_Grow(c *mcentral) *mspan { |
| npages := uintptr(class_to_allocnpages[c.sizeclass]) |
| size := uintptr(class_to_size[c.sizeclass]) |
| n := (npages << _PageShift) / size |
| |
| s := mHeap_Alloc(&mheap_, npages, c.sizeclass, false, true) |
| if s == nil { |
| return nil |
| } |
| |
| p := uintptr(s.start << _PageShift) |
| s.limit = p + size*n |
| head := gclinkptr(p) |
| tail := gclinkptr(p) |
| // i==0 iteration already done |
| for i := uintptr(1); i < n; i++ { |
| p += size |
| tail.ptr().next = gclinkptr(p) |
| tail = gclinkptr(p) |
| } |
| if s.freelist.ptr() != nil { |
| throw("freelist not empty") |
| } |
| tail.ptr().next = 0 |
| s.freelist = head |
| heapBitsForSpan(s.base()).initSpan(s.layout()) |
| return s |
| } |