| // Copyright 2011 The Go Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style |
| // license that can be found in the LICENSE file. |
| |
| package runtime |
| |
| import ( |
| "internal/abi" |
| "runtime/internal/atomic" |
| "unsafe" |
| ) |
| |
| type mOS struct { |
| waitsemacount uint32 |
| } |
| |
| const ( |
| _ESRCH = 3 |
| _EWOULDBLOCK = _EAGAIN |
| _ENOTSUP = 91 |
| |
| // From OpenBSD's sys/time.h |
| _CLOCK_REALTIME = 0 |
| _CLOCK_VIRTUAL = 1 |
| _CLOCK_PROF = 2 |
| _CLOCK_MONOTONIC = 3 |
| ) |
| |
| type sigset uint32 |
| |
| var sigset_all = ^sigset(0) |
| |
| // From OpenBSD's <sys/sysctl.h> |
| const ( |
| _CTL_HW = 6 |
| _HW_NCPU = 3 |
| _HW_PAGESIZE = 7 |
| _HW_NCPUONLINE = 25 |
| ) |
| |
| func sysctlInt(mib []uint32) (int32, bool) { |
| var out int32 |
| nout := unsafe.Sizeof(out) |
| ret := sysctl(&mib[0], uint32(len(mib)), (*byte)(unsafe.Pointer(&out)), &nout, nil, 0) |
| if ret < 0 { |
| return 0, false |
| } |
| return out, true |
| } |
| |
| func sysctlUint64(mib []uint32) (uint64, bool) { |
| var out uint64 |
| nout := unsafe.Sizeof(out) |
| ret := sysctl(&mib[0], uint32(len(mib)), (*byte)(unsafe.Pointer(&out)), &nout, nil, 0) |
| if ret < 0 { |
| return 0, false |
| } |
| return out, true |
| } |
| |
| //go:linkname internal_cpu_sysctlUint64 internal/cpu.sysctlUint64 |
| func internal_cpu_sysctlUint64(mib []uint32) (uint64, bool) { |
| return sysctlUint64(mib) |
| } |
| |
| func getncpu() int32 { |
| // Try hw.ncpuonline first because hw.ncpu would report a number twice as |
| // high as the actual CPUs running on OpenBSD 6.4 with hyperthreading |
| // disabled (hw.smt=0). See https://golang.org/issue/30127 |
| if n, ok := sysctlInt([]uint32{_CTL_HW, _HW_NCPUONLINE}); ok { |
| return int32(n) |
| } |
| if n, ok := sysctlInt([]uint32{_CTL_HW, _HW_NCPU}); ok { |
| return int32(n) |
| } |
| return 1 |
| } |
| |
| func getPageSize() uintptr { |
| if ps, ok := sysctlInt([]uint32{_CTL_HW, _HW_PAGESIZE}); ok { |
| return uintptr(ps) |
| } |
| return 0 |
| } |
| |
| //go:nosplit |
| func semacreate(mp *m) { |
| } |
| |
| //go:nosplit |
| func semasleep(ns int64) int32 { |
| gp := getg() |
| |
| // Compute sleep deadline. |
| var tsp *timespec |
| if ns >= 0 { |
| var ts timespec |
| ts.setNsec(ns + nanotime()) |
| tsp = &ts |
| } |
| |
| for { |
| v := atomic.Load(&gp.m.waitsemacount) |
| if v > 0 { |
| if atomic.Cas(&gp.m.waitsemacount, v, v-1) { |
| return 0 // semaphore acquired |
| } |
| continue |
| } |
| |
| // Sleep until woken by semawakeup or timeout; or abort if waitsemacount != 0. |
| // |
| // From OpenBSD's __thrsleep(2) manual: |
| // "The abort argument, if not NULL, points to an int that will |
| // be examined [...] immediately before blocking. If that int |
| // is non-zero then __thrsleep() will immediately return EINTR |
| // without blocking." |
| ret := thrsleep(uintptr(unsafe.Pointer(&gp.m.waitsemacount)), _CLOCK_MONOTONIC, tsp, 0, &gp.m.waitsemacount) |
| if ret == _EWOULDBLOCK { |
| return -1 |
| } |
| } |
| } |
| |
| //go:nosplit |
| func semawakeup(mp *m) { |
| atomic.Xadd(&mp.waitsemacount, 1) |
| ret := thrwakeup(uintptr(unsafe.Pointer(&mp.waitsemacount)), 1) |
| if ret != 0 && ret != _ESRCH { |
| // semawakeup can be called on signal stack. |
| systemstack(func() { |
| print("thrwakeup addr=", &mp.waitsemacount, " sem=", mp.waitsemacount, " ret=", ret, "\n") |
| }) |
| } |
| } |
| |
| func osinit() { |
| ncpu = getncpu() |
| physPageSize = getPageSize() |
| } |
| |
| var urandom_dev = []byte("/dev/urandom\x00") |
| |
| //go:nosplit |
| func readRandom(r []byte) int { |
| fd := open(&urandom_dev[0], 0 /* O_RDONLY */, 0) |
| n := read(fd, unsafe.Pointer(&r[0]), int32(len(r))) |
| closefd(fd) |
| return int(n) |
| } |
| |
| func goenvs() { |
| goenvs_unix() |
| } |
| |
| // Called to initialize a new m (including the bootstrap m). |
| // Called on the parent thread (main thread in case of bootstrap), can allocate memory. |
| func mpreinit(mp *m) { |
| gsignalSize := int32(32 * 1024) |
| if GOARCH == "mips64" { |
| gsignalSize = int32(64 * 1024) |
| } |
| mp.gsignal = malg(gsignalSize) |
| mp.gsignal.m = mp |
| } |
| |
| // Called to initialize a new m (including the bootstrap m). |
| // Called on the new thread, can not allocate memory. |
| func minit() { |
| getg().m.procid = uint64(getthrid()) |
| minitSignals() |
| } |
| |
| // Called from dropm to undo the effect of an minit. |
| // |
| //go:nosplit |
| func unminit() { |
| unminitSignals() |
| getg().m.procid = 0 |
| } |
| |
| // Called from exitm, but not from drop, to undo the effect of thread-owned |
| // resources in minit, semacreate, or elsewhere. Do not take locks after calling this. |
| func mdestroy(mp *m) { |
| } |
| |
| func sigtramp() |
| |
| type sigactiont struct { |
| sa_sigaction uintptr |
| sa_mask uint32 |
| sa_flags int32 |
| } |
| |
| //go:nosplit |
| //go:nowritebarrierrec |
| func setsig(i uint32, fn uintptr) { |
| var sa sigactiont |
| sa.sa_flags = _SA_SIGINFO | _SA_ONSTACK | _SA_RESTART |
| sa.sa_mask = uint32(sigset_all) |
| if fn == abi.FuncPCABIInternal(sighandler) { // abi.FuncPCABIInternal(sighandler) matches the callers in signal_unix.go |
| fn = abi.FuncPCABI0(sigtramp) |
| } |
| sa.sa_sigaction = fn |
| sigaction(i, &sa, nil) |
| } |
| |
| //go:nosplit |
| //go:nowritebarrierrec |
| func setsigstack(i uint32) { |
| throw("setsigstack") |
| } |
| |
| //go:nosplit |
| //go:nowritebarrierrec |
| func getsig(i uint32) uintptr { |
| var sa sigactiont |
| sigaction(i, nil, &sa) |
| return sa.sa_sigaction |
| } |
| |
| // setSignalstackSP sets the ss_sp field of a stackt. |
| // |
| //go:nosplit |
| func setSignalstackSP(s *stackt, sp uintptr) { |
| s.ss_sp = sp |
| } |
| |
| //go:nosplit |
| //go:nowritebarrierrec |
| func sigaddset(mask *sigset, i int) { |
| *mask |= 1 << (uint32(i) - 1) |
| } |
| |
| func sigdelset(mask *sigset, i int) { |
| *mask &^= 1 << (uint32(i) - 1) |
| } |
| |
| //go:nosplit |
| func (c *sigctxt) fixsigcode(sig uint32) { |
| } |
| |
| func setProcessCPUProfiler(hz int32) { |
| setProcessCPUProfilerTimer(hz) |
| } |
| |
| func setThreadCPUProfiler(hz int32) { |
| setThreadCPUProfilerHz(hz) |
| } |
| |
| //go:nosplit |
| func validSIGPROF(mp *m, c *sigctxt) bool { |
| return true |
| } |
| |
| func osStackAlloc(s *mspan) { |
| osStackRemap(s, _MAP_STACK) |
| } |
| |
| func osStackFree(s *mspan) { |
| // Undo MAP_STACK. |
| osStackRemap(s, 0) |
| } |
| |
| func osStackRemap(s *mspan, flags int32) { |
| a, err := mmap(unsafe.Pointer(s.base()), s.npages*pageSize, _PROT_READ|_PROT_WRITE, _MAP_PRIVATE|_MAP_ANON|_MAP_FIXED|flags, -1, 0) |
| if err != 0 || uintptr(a) != s.base() { |
| print("runtime: remapping stack memory ", hex(s.base()), " ", s.npages*pageSize, " a=", a, " err=", err, "\n") |
| throw("remapping stack memory failed") |
| } |
| } |
| |
| //go:nosplit |
| func raise(sig uint32) { |
| thrkill(getthrid(), int(sig)) |
| } |
| |
| func signalM(mp *m, sig int) { |
| thrkill(int32(mp.procid), sig) |
| } |
| |
| // sigPerThreadSyscall is only used on linux, so we assign a bogus signal |
| // number. |
| const sigPerThreadSyscall = 1 << 31 |
| |
| //go:nosplit |
| func runPerThreadSyscall() { |
| throw("runPerThreadSyscall only valid on linux") |
| } |