| // Copyright 2009 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| // Memory statistics |
| |
| package runtime |
| |
| import ( |
| "runtime/internal/atomic" |
| "unsafe" |
| ) |
| |
| // Statistics. |
| // |
| // For detailed descriptions see the documentation for MemStats. |
| // Fields that differ from MemStats are further documented here. |
| // |
| // Many of these fields are updated on the fly, while others are only |
| // updated when updatememstats is called. |
| type mstats struct { |
| // General statistics. |
| alloc uint64 // bytes allocated and not yet freed |
| total_alloc uint64 // bytes allocated (even if freed) |
| sys uint64 // bytes obtained from system (should be sum of xxx_sys below, no locking, approximate) |
| nlookup uint64 // number of pointer lookups (unused) |
| nmalloc uint64 // number of mallocs |
| nfree uint64 // number of frees |
| |
| // Statistics about malloc heap. |
| // Updated atomically, or with the world stopped. |
| // |
| // Like MemStats, heap_sys and heap_inuse do not count memory |
| // in manually-managed spans. |
| heap_sys sysMemStat // virtual address space obtained from system for GC'd heap |
| heap_inuse uint64 // bytes in mSpanInUse spans |
| heap_released uint64 // bytes released to the os |
| |
| // heap_objects is not used by the runtime directly and instead |
| // computed on the fly by updatememstats. |
| heap_objects uint64 // total number of allocated objects |
| |
| // Statistics about allocation of low-level fixed-size structures. |
| // Protected by FixAlloc locks. |
| stacks_inuse uint64 // bytes in manually-managed stack spans; updated atomically or during STW |
| stacks_sys sysMemStat // only counts newosproc0 stack in mstats; differs from MemStats.StackSys |
| mspan_inuse uint64 // mspan structures |
| mspan_sys sysMemStat |
| mcache_inuse uint64 // mcache structures |
| mcache_sys sysMemStat |
| buckhash_sys sysMemStat // profiling bucket hash table |
| gcWorkBufInUse uint64 // updated atomically or during STW |
| gcProgPtrScalarBitsInUse uint64 // updated atomically or during STW |
| gcMiscSys sysMemStat // updated atomically or during STW |
| other_sys sysMemStat // updated atomically or during STW |
| |
| // Statistics about the garbage collector. |
| |
| // next_gc is the goal heap_live for when next GC ends. |
| // Set to ^uint64(0) if disabled. |
| // |
| // Read and written atomically, unless the world is stopped. |
| next_gc uint64 |
| |
| // Protected by mheap or stopping the world during GC. |
| last_gc_unix uint64 // last gc (in unix time) |
| pause_total_ns uint64 |
| pause_ns [256]uint64 // circular buffer of recent gc pause lengths |
| pause_end [256]uint64 // circular buffer of recent gc end times (nanoseconds since 1970) |
| numgc uint32 |
| numforcedgc uint32 // number of user-forced GCs |
| gc_cpu_fraction float64 // fraction of CPU time used by GC |
| enablegc bool |
| debuggc bool |
| |
| // Statistics about allocation size classes. |
| |
| by_size [_NumSizeClasses]struct { |
| size uint32 |
| nmalloc uint64 |
| nfree uint64 |
| } |
| |
| // Add an uint32 for even number of size classes to align below fields |
| // to 64 bits for atomic operations on 32 bit platforms. |
| _ [1 - _NumSizeClasses%2]uint32 |
| |
| last_gc_nanotime uint64 // last gc (monotonic time) |
| tinyallocs uint64 // number of tiny allocations that didn't cause actual allocation; not exported to go directly |
| last_next_gc uint64 // next_gc for the previous GC |
| last_heap_inuse uint64 // heap_inuse at mark termination of the previous GC |
| |
| // triggerRatio is the heap growth ratio that triggers marking. |
| // |
| // E.g., if this is 0.6, then GC should start when the live |
| // heap has reached 1.6 times the heap size marked by the |
| // previous cycle. This should be ≤ GOGC/100 so the trigger |
| // heap size is less than the goal heap size. This is set |
| // during mark termination for the next cycle's trigger. |
| triggerRatio float64 |
| |
| // gc_trigger is the heap size that triggers marking. |
| // |
| // When heap_live ≥ gc_trigger, the mark phase will start. |
| // This is also the heap size by which proportional sweeping |
| // must be complete. |
| // |
| // This is computed from triggerRatio during mark termination |
| // for the next cycle's trigger. |
| gc_trigger uint64 |
| |
| // heap_live is the number of bytes considered live by the GC. |
| // That is: retained by the most recent GC plus allocated |
| // since then. heap_live <= alloc, since alloc includes unmarked |
| // objects that have not yet been swept (and hence goes up as we |
| // allocate and down as we sweep) while heap_live excludes these |
| // objects (and hence only goes up between GCs). |
| // |
| // This is updated atomically without locking. To reduce |
| // contention, this is updated only when obtaining a span from |
| // an mcentral and at this point it counts all of the |
| // unallocated slots in that span (which will be allocated |
| // before that mcache obtains another span from that |
| // mcentral). Hence, it slightly overestimates the "true" live |
| // heap size. It's better to overestimate than to |
| // underestimate because 1) this triggers the GC earlier than |
| // necessary rather than potentially too late and 2) this |
| // leads to a conservative GC rate rather than a GC rate that |
| // is potentially too low. |
| // |
| // Reads should likewise be atomic (or during STW). |
| // |
| // Whenever this is updated, call traceHeapAlloc() and |
| // gcController.revise(). |
| heap_live uint64 |
| |
| // heap_scan is the number of bytes of "scannable" heap. This |
| // is the live heap (as counted by heap_live), but omitting |
| // no-scan objects and no-scan tails of objects. |
| // |
| // Whenever this is updated, call gcController.revise(). |
| // |
| // Read and written atomically or with the world stopped. |
| heap_scan uint64 |
| |
| // heap_marked is the number of bytes marked by the previous |
| // GC. After mark termination, heap_live == heap_marked, but |
| // unlike heap_live, heap_marked does not change until the |
| // next mark termination. |
| heap_marked uint64 |
| } |
| |
| var memstats mstats |
| |
| // A MemStats records statistics about the memory allocator. |
| type MemStats struct { |
| // General statistics. |
| |
| // Alloc is bytes of allocated heap objects. |
| // |
| // This is the same as HeapAlloc (see below). |
| Alloc uint64 |
| |
| // TotalAlloc is cumulative bytes allocated for heap objects. |
| // |
| // TotalAlloc increases as heap objects are allocated, but |
| // unlike Alloc and HeapAlloc, it does not decrease when |
| // objects are freed. |
| TotalAlloc uint64 |
| |
| // Sys is the total bytes of memory obtained from the OS. |
| // |
| // Sys is the sum of the XSys fields below. Sys measures the |
| // virtual address space reserved by the Go runtime for the |
| // heap, stacks, and other internal data structures. It's |
| // likely that not all of the virtual address space is backed |
| // by physical memory at any given moment, though in general |
| // it all was at some point. |
| Sys uint64 |
| |
| // Lookups is the number of pointer lookups performed by the |
| // runtime. |
| // |
| // This is primarily useful for debugging runtime internals. |
| Lookups uint64 |
| |
| // Mallocs is the cumulative count of heap objects allocated. |
| // The number of live objects is Mallocs - Frees. |
| Mallocs uint64 |
| |
| // Frees is the cumulative count of heap objects freed. |
| Frees uint64 |
| |
| // Heap memory statistics. |
| // |
| // Interpreting the heap statistics requires some knowledge of |
| // how Go organizes memory. Go divides the virtual address |
| // space of the heap into "spans", which are contiguous |
| // regions of memory 8K or larger. A span may be in one of |
| // three states: |
| // |
| // An "idle" span contains no objects or other data. The |
| // physical memory backing an idle span can be released back |
| // to the OS (but the virtual address space never is), or it |
| // can be converted into an "in use" or "stack" span. |
| // |
| // An "in use" span contains at least one heap object and may |
| // have free space available to allocate more heap objects. |
| // |
| // A "stack" span is used for goroutine stacks. Stack spans |
| // are not considered part of the heap. A span can change |
| // between heap and stack memory; it is never used for both |
| // simultaneously. |
| |
| // HeapAlloc is bytes of allocated heap objects. |
| // |
| // "Allocated" heap objects include all reachable objects, as |
| // well as unreachable objects that the garbage collector has |
| // not yet freed. Specifically, HeapAlloc increases as heap |
| // objects are allocated and decreases as the heap is swept |
| // and unreachable objects are freed. Sweeping occurs |
| // incrementally between GC cycles, so these two processes |
| // occur simultaneously, and as a result HeapAlloc tends to |
| // change smoothly (in contrast with the sawtooth that is |
| // typical of stop-the-world garbage collectors). |
| HeapAlloc uint64 |
| |
| // HeapSys is bytes of heap memory obtained from the OS. |
| // |
| // HeapSys measures the amount of virtual address space |
| // reserved for the heap. This includes virtual address space |
| // that has been reserved but not yet used, which consumes no |
| // physical memory, but tends to be small, as well as virtual |
| // address space for which the physical memory has been |
| // returned to the OS after it became unused (see HeapReleased |
| // for a measure of the latter). |
| // |
| // HeapSys estimates the largest size the heap has had. |
| HeapSys uint64 |
| |
| // HeapIdle is bytes in idle (unused) spans. |
| // |
| // Idle spans have no objects in them. These spans could be |
| // (and may already have been) returned to the OS, or they can |
| // be reused for heap allocations, or they can be reused as |
| // stack memory. |
| // |
| // HeapIdle minus HeapReleased estimates the amount of memory |
| // that could be returned to the OS, but is being retained by |
| // the runtime so it can grow the heap without requesting more |
| // memory from the OS. If this difference is significantly |
| // larger than the heap size, it indicates there was a recent |
| // transient spike in live heap size. |
| HeapIdle uint64 |
| |
| // HeapInuse is bytes in in-use spans. |
| // |
| // In-use spans have at least one object in them. These spans |
| // can only be used for other objects of roughly the same |
| // size. |
| // |
| // HeapInuse minus HeapAlloc estimates the amount of memory |
| // that has been dedicated to particular size classes, but is |
| // not currently being used. This is an upper bound on |
| // fragmentation, but in general this memory can be reused |
| // efficiently. |
| HeapInuse uint64 |
| |
| // HeapReleased is bytes of physical memory returned to the OS. |
| // |
| // This counts heap memory from idle spans that was returned |
| // to the OS and has not yet been reacquired for the heap. |
| HeapReleased uint64 |
| |
| // HeapObjects is the number of allocated heap objects. |
| // |
| // Like HeapAlloc, this increases as objects are allocated and |
| // decreases as the heap is swept and unreachable objects are |
| // freed. |
| HeapObjects uint64 |
| |
| // Stack memory statistics. |
| // |
| // Stacks are not considered part of the heap, but the runtime |
| // can reuse a span of heap memory for stack memory, and |
| // vice-versa. |
| |
| // StackInuse is bytes in stack spans. |
| // |
| // In-use stack spans have at least one stack in them. These |
| // spans can only be used for other stacks of the same size. |
| // |
| // There is no StackIdle because unused stack spans are |
| // returned to the heap (and hence counted toward HeapIdle). |
| StackInuse uint64 |
| |
| // StackSys is bytes of stack memory obtained from the OS. |
| // |
| // StackSys is StackInuse, plus any memory obtained directly |
| // from the OS for OS thread stacks (which should be minimal). |
| StackSys uint64 |
| |
| // Off-heap memory statistics. |
| // |
| // The following statistics measure runtime-internal |
| // structures that are not allocated from heap memory (usually |
| // because they are part of implementing the heap). Unlike |
| // heap or stack memory, any memory allocated to these |
| // structures is dedicated to these structures. |
| // |
| // These are primarily useful for debugging runtime memory |
| // overheads. |
| |
| // MSpanInuse is bytes of allocated mspan structures. |
| MSpanInuse uint64 |
| |
| // MSpanSys is bytes of memory obtained from the OS for mspan |
| // structures. |
| MSpanSys uint64 |
| |
| // MCacheInuse is bytes of allocated mcache structures. |
| MCacheInuse uint64 |
| |
| // MCacheSys is bytes of memory obtained from the OS for |
| // mcache structures. |
| MCacheSys uint64 |
| |
| // BuckHashSys is bytes of memory in profiling bucket hash tables. |
| BuckHashSys uint64 |
| |
| // GCSys is bytes of memory in garbage collection metadata. |
| GCSys uint64 |
| |
| // OtherSys is bytes of memory in miscellaneous off-heap |
| // runtime allocations. |
| OtherSys uint64 |
| |
| // Garbage collector statistics. |
| |
| // NextGC is the target heap size of the next GC cycle. |
| // |
| // The garbage collector's goal is to keep HeapAlloc ≤ NextGC. |
| // At the end of each GC cycle, the target for the next cycle |
| // is computed based on the amount of reachable data and the |
| // value of GOGC. |
| NextGC uint64 |
| |
| // LastGC is the time the last garbage collection finished, as |
| // nanoseconds since 1970 (the UNIX epoch). |
| LastGC uint64 |
| |
| // PauseTotalNs is the cumulative nanoseconds in GC |
| // stop-the-world pauses since the program started. |
| // |
| // During a stop-the-world pause, all goroutines are paused |
| // and only the garbage collector can run. |
| PauseTotalNs uint64 |
| |
| // PauseNs is a circular buffer of recent GC stop-the-world |
| // pause times in nanoseconds. |
| // |
| // The most recent pause is at PauseNs[(NumGC+255)%256]. In |
| // general, PauseNs[N%256] records the time paused in the most |
| // recent N%256th GC cycle. There may be multiple pauses per |
| // GC cycle; this is the sum of all pauses during a cycle. |
| PauseNs [256]uint64 |
| |
| // PauseEnd is a circular buffer of recent GC pause end times, |
| // as nanoseconds since 1970 (the UNIX epoch). |
| // |
| // This buffer is filled the same way as PauseNs. There may be |
| // multiple pauses per GC cycle; this records the end of the |
| // last pause in a cycle. |
| PauseEnd [256]uint64 |
| |
| // NumGC is the number of completed GC cycles. |
| NumGC uint32 |
| |
| // NumForcedGC is the number of GC cycles that were forced by |
| // the application calling the GC function. |
| NumForcedGC uint32 |
| |
| // GCCPUFraction is the fraction of this program's available |
| // CPU time used by the GC since the program started. |
| // |
| // GCCPUFraction is expressed as a number between 0 and 1, |
| // where 0 means GC has consumed none of this program's CPU. A |
| // program's available CPU time is defined as the integral of |
| // GOMAXPROCS since the program started. That is, if |
| // GOMAXPROCS is 2 and a program has been running for 10 |
| // seconds, its "available CPU" is 20 seconds. GCCPUFraction |
| // does not include CPU time used for write barrier activity. |
| // |
| // This is the same as the fraction of CPU reported by |
| // GODEBUG=gctrace=1. |
| GCCPUFraction float64 |
| |
| // EnableGC indicates that GC is enabled. It is always true, |
| // even if GOGC=off. |
| EnableGC bool |
| |
| // DebugGC is currently unused. |
| DebugGC bool |
| |
| // BySize reports per-size class allocation statistics. |
| // |
| // BySize[N] gives statistics for allocations of size S where |
| // BySize[N-1].Size < S ≤ BySize[N].Size. |
| // |
| // This does not report allocations larger than BySize[60].Size. |
| BySize [61]struct { |
| // Size is the maximum byte size of an object in this |
| // size class. |
| Size uint32 |
| |
| // Mallocs is the cumulative count of heap objects |
| // allocated in this size class. The cumulative bytes |
| // of allocation is Size*Mallocs. The number of live |
| // objects in this size class is Mallocs - Frees. |
| Mallocs uint64 |
| |
| // Frees is the cumulative count of heap objects freed |
| // in this size class. |
| Frees uint64 |
| } |
| } |
| |
| func init() { |
| if unsafe.Offsetof(memstats.heap_live)%8 != 0 { |
| println(unsafe.Offsetof(memstats.heap_live)) |
| throw("memstats.heap_live not aligned to 8 bytes") |
| } |
| } |
| |
| // ReadMemStats populates m with memory allocator statistics. |
| // |
| // The returned memory allocator statistics are up to date as of the |
| // call to ReadMemStats. This is in contrast with a heap profile, |
| // which is a snapshot as of the most recently completed garbage |
| // collection cycle. |
| func ReadMemStats(m *MemStats) { |
| stopTheWorld("read mem stats") |
| |
| systemstack(func() { |
| readmemstats_m(m) |
| }) |
| |
| startTheWorld() |
| } |
| |
| func readmemstats_m(stats *MemStats) { |
| updatememstats() |
| |
| stats.Alloc = memstats.alloc |
| stats.TotalAlloc = memstats.total_alloc |
| stats.Sys = memstats.sys |
| stats.Mallocs = memstats.nmalloc |
| stats.Frees = memstats.nfree |
| stats.HeapAlloc = memstats.alloc |
| stats.HeapSys = memstats.heap_sys.load() |
| // By definition, HeapIdle is memory that was mapped |
| // for the heap but is not currently used to hold heap |
| // objects. It also specifically is memory that can be |
| // used for other purposes, like stacks, but this memory |
| // is subtracted out of HeapSys before it makes that |
| // transition. Put another way: |
| // |
| // heap_sys = bytes allocated from the OS for the heap - bytes ultimately used for non-heap purposes |
| // heap_idle = bytes allocated from the OS for the heap - bytes ultimately used for any purpose |
| // |
| // or |
| // |
| // heap_sys = sys - stacks_inuse - gcWorkBufInUse - gcProgPtrScalarBitsInUse |
| // heap_idle = sys - stacks_inuse - gcWorkBufInUse - gcProgPtrScalarBitsInUse - heap_inuse |
| // |
| // => heap_idle = heap_sys - heap_inuse |
| stats.HeapIdle = memstats.heap_sys.load() - memstats.heap_inuse |
| stats.HeapInuse = memstats.heap_inuse |
| stats.HeapReleased = memstats.heap_released |
| stats.HeapObjects = memstats.heap_objects |
| stats.StackInuse = memstats.stacks_inuse |
| // memstats.stacks_sys is only memory mapped directly for OS stacks. |
| // Add in heap-allocated stack memory for user consumption. |
| stats.StackSys = memstats.stacks_inuse + memstats.stacks_sys.load() |
| stats.MSpanInuse = memstats.mspan_inuse |
| stats.MSpanSys = memstats.mspan_sys.load() |
| stats.MCacheInuse = memstats.mcache_inuse |
| stats.MCacheSys = memstats.mcache_sys.load() |
| stats.BuckHashSys = memstats.buckhash_sys.load() |
| // MemStats defines GCSys as an aggregate of all memory related |
| // to the memory management system, but we track this memory |
| // at a more granular level in the runtime. |
| stats.GCSys = memstats.gcMiscSys.load() + memstats.gcWorkBufInUse + memstats.gcProgPtrScalarBitsInUse |
| stats.OtherSys = memstats.other_sys.load() |
| stats.NextGC = memstats.next_gc |
| stats.LastGC = memstats.last_gc_unix |
| stats.PauseTotalNs = memstats.pause_total_ns |
| stats.PauseNs = memstats.pause_ns |
| stats.PauseEnd = memstats.pause_end |
| stats.NumGC = memstats.numgc |
| stats.NumForcedGC = memstats.numforcedgc |
| stats.GCCPUFraction = memstats.gc_cpu_fraction |
| stats.EnableGC = true |
| |
| // Handle BySize. Copy N values, where N is |
| // the minimum of the lengths of the two arrays. |
| // Unfortunately copy() won't work here because |
| // the arrays have different structs. |
| // |
| // TODO(mknyszek): Consider renaming the fields |
| // of by_size's elements to align so we can use |
| // the copy built-in. |
| bySizeLen := len(stats.BySize) |
| if l := len(memstats.by_size); l < bySizeLen { |
| bySizeLen = l |
| } |
| for i := 0; i < bySizeLen; i++ { |
| stats.BySize[i].Size = memstats.by_size[i].size |
| stats.BySize[i].Mallocs = memstats.by_size[i].nmalloc |
| stats.BySize[i].Frees = memstats.by_size[i].nfree |
| } |
| } |
| |
| //go:linkname readGCStats runtime/debug.readGCStats |
| func readGCStats(pauses *[]uint64) { |
| systemstack(func() { |
| readGCStats_m(pauses) |
| }) |
| } |
| |
| // readGCStats_m must be called on the system stack because it acquires the heap |
| // lock. See mheap for details. |
| //go:systemstack |
| func readGCStats_m(pauses *[]uint64) { |
| p := *pauses |
| // Calling code in runtime/debug should make the slice large enough. |
| if cap(p) < len(memstats.pause_ns)+3 { |
| throw("short slice passed to readGCStats") |
| } |
| |
| // Pass back: pauses, pause ends, last gc (absolute time), number of gc, total pause ns. |
| lock(&mheap_.lock) |
| |
| n := memstats.numgc |
| if n > uint32(len(memstats.pause_ns)) { |
| n = uint32(len(memstats.pause_ns)) |
| } |
| |
| // The pause buffer is circular. The most recent pause is at |
| // pause_ns[(numgc-1)%len(pause_ns)], and then backward |
| // from there to go back farther in time. We deliver the times |
| // most recent first (in p[0]). |
| p = p[:cap(p)] |
| for i := uint32(0); i < n; i++ { |
| j := (memstats.numgc - 1 - i) % uint32(len(memstats.pause_ns)) |
| p[i] = memstats.pause_ns[j] |
| p[n+i] = memstats.pause_end[j] |
| } |
| |
| p[n+n] = memstats.last_gc_unix |
| p[n+n+1] = uint64(memstats.numgc) |
| p[n+n+2] = memstats.pause_total_ns |
| unlock(&mheap_.lock) |
| *pauses = p[:n+n+3] |
| } |
| |
| //go:nowritebarrier |
| func updatememstats() { |
| // Flush mcaches to mcentral before doing anything else. |
| // |
| // Flushing to the mcentral may in general cause stats to |
| // change as mcentral data structures are manipulated. |
| systemstack(flushallmcaches) |
| |
| memstats.mcache_inuse = uint64(mheap_.cachealloc.inuse) |
| memstats.mspan_inuse = uint64(mheap_.spanalloc.inuse) |
| memstats.sys = memstats.heap_sys.load() + memstats.stacks_sys.load() + memstats.mspan_sys.load() + |
| memstats.mcache_sys.load() + memstats.buckhash_sys.load() + memstats.gcMiscSys.load() + |
| memstats.other_sys.load() |
| |
| // We also count stacks_inuse, gcWorkBufInUse, and gcProgPtrScalarBitsInUse as sys memory. |
| memstats.sys += memstats.stacks_inuse + memstats.gcWorkBufInUse + memstats.gcProgPtrScalarBitsInUse |
| |
| // Calculate memory allocator stats. |
| // During program execution we only count number of frees and amount of freed memory. |
| // Current number of alive objects in the heap and amount of alive heap memory |
| // are calculated by scanning all spans. |
| // Total number of mallocs is calculated as number of frees plus number of alive objects. |
| // Similarly, total amount of allocated memory is calculated as amount of freed memory |
| // plus amount of alive heap memory. |
| memstats.alloc = 0 |
| memstats.total_alloc = 0 |
| memstats.nmalloc = 0 |
| memstats.nfree = 0 |
| memstats.tinyallocs = 0 |
| for i := 0; i < len(memstats.by_size); i++ { |
| memstats.by_size[i].nmalloc = 0 |
| memstats.by_size[i].nfree = 0 |
| } |
| |
| // Collect allocation stats. This is safe and consistent |
| // because the world is stopped. |
| var smallFree, totalAlloc, totalFree uint64 |
| for _, p := range allp { |
| c := p.mcache |
| if c == nil { |
| continue |
| } |
| // Collect large allocation stats. |
| memstats.nmalloc += uint64(c.largeAllocCount) |
| totalAlloc += uint64(c.largeAlloc) |
| totalFree += uint64(c.largeFree) |
| memstats.nfree += uint64(c.largeFreeCount) |
| |
| // Collect tiny allocation stats. |
| memstats.tinyallocs += uint64(c.tinyAllocCount) |
| |
| // Collect per-sizeclass stats. |
| for i := 0; i < _NumSizeClasses; i++ { |
| // Malloc stats. |
| memstats.nmalloc += uint64(c.smallAllocCount[i]) |
| memstats.by_size[i].nmalloc += uint64(c.smallAllocCount[i]) |
| totalAlloc += uint64(c.smallAllocCount[i]) * uint64(class_to_size[i]) |
| |
| // Free stats. |
| memstats.nfree += uint64(c.smallFreeCount[i]) |
| memstats.by_size[i].nfree += uint64(c.smallFreeCount[i]) |
| smallFree += uint64(c.smallFreeCount[i]) * uint64(class_to_size[i]) |
| } |
| } |
| |
| totalFree += smallFree |
| |
| memstats.nfree += memstats.tinyallocs |
| memstats.nmalloc += memstats.tinyallocs |
| |
| // Calculate derived stats. |
| memstats.total_alloc = totalAlloc |
| memstats.alloc = totalAlloc - totalFree |
| memstats.heap_objects = memstats.nmalloc - memstats.nfree |
| } |
| |
| // flushmcache flushes the mcache of allp[i]. |
| // |
| // The world must be stopped. |
| // |
| //go:nowritebarrier |
| func flushmcache(i int) { |
| p := allp[i] |
| c := p.mcache |
| if c == nil { |
| return |
| } |
| c.releaseAll() |
| stackcache_clear(c) |
| } |
| |
| // flushallmcaches flushes the mcaches of all Ps. |
| // |
| // The world must be stopped. |
| // |
| //go:nowritebarrier |
| func flushallmcaches() { |
| for i := 0; i < int(gomaxprocs); i++ { |
| flushmcache(i) |
| } |
| } |
| |
| // sysMemStat represents a global system statistic that is managed atomically. |
| // |
| // This type must structurally be a uint64 so that mstats aligns with MemStats. |
| type sysMemStat uint64 |
| |
| // load atomically reads the value of the stat. |
| func (s *sysMemStat) load() uint64 { |
| return atomic.Load64((*uint64)(s)) |
| } |
| |
| // add atomically adds the sysMemStat by n. |
| func (s *sysMemStat) add(n int64) { |
| if s == nil { |
| return |
| } |
| val := atomic.Xadd64((*uint64)(s), n) |
| if (n > 0 && int64(val) < n) || (n < 0 && int64(val)+n < n) { |
| print("runtime: val=", val, " n=", n, "\n") |
| throw("sysMemStat overflow") |
| } |
| } |