Source File
cpuprof.go
Belonging Package
runtime
// Copyright 2011 The Go Authors. All rights reserved.// Use of this source code is governed by a BSD-style// license that can be found in the LICENSE file.// CPU profiling.//// The signal handler for the profiling clock tick adds a new stack trace// to a log of recent traces. The log is read by a user goroutine that// turns it into formatted profile data. If the reader does not keep up// with the log, those writes will be recorded as a count of lost records.// The actual profile buffer is in profbuf.go.package runtimeimport ()const (maxCPUProfStack = 64// profBufWordCount is the size of the CPU profile buffer's storage for the// header and stack of each sample, measured in 64-bit words. Every sample// has a required header of two words. With a small additional header (a// word or two) and stacks at the profiler's maximum length of 64 frames,// that capacity can support 1900 samples or 19 thread-seconds at a 100 Hz// sample rate, at a cost of 1 MiB.profBufWordCount = 1 << 17// profBufTagCount is the size of the CPU profile buffer's storage for the// goroutine tags associated with each sample. A capacity of 1<<14 means// room for 16k samples, or 160 thread-seconds at a 100 Hz sample rate.profBufTagCount = 1 << 14)type cpuProfile struct {lock mutexon bool // profiling is onlog *profBuf // profile events written here// extra holds extra stacks accumulated in addNonGo// corresponding to profiling signals arriving on// non-Go-created threads. Those stacks are written// to log the next time a normal Go thread gets the// signal handler.// Assuming the stacks are 2 words each (we don't get// a full traceback from those threads), plus one word// size for framing, 100 Hz profiling would generate// 300 words per second.// Hopefully a normal Go thread will get the profiling// signal at least once every few seconds.extra [1000]uintptrnumExtra intlostExtra uint64 // count of frames lost because extra is fulllostAtomic uint64 // count of frames lost because of being in atomic64 on mips/arm; updated racily}var cpuprof cpuProfile// SetCPUProfileRate sets the CPU profiling rate to hz samples per second.// If hz <= 0, SetCPUProfileRate turns off profiling.// If the profiler is on, the rate cannot be changed without first turning it off.//// Most clients should use the runtime/pprof package or// the testing package's -test.cpuprofile flag instead of calling// SetCPUProfileRate directly.func ( int) {// Clamp hz to something reasonable.if < 0 {= 0}if > 1000000 {= 1000000}lock(&cpuprof.lock)if > 0 {if cpuprof.on || cpuprof.log != nil {print("runtime: cannot set cpu profile rate until previous profile has finished.\n")unlock(&cpuprof.lock)return}cpuprof.on = truecpuprof.log = newProfBuf(1, profBufWordCount, profBufTagCount):= [1]uint64{uint64()}cpuprof.log.write(nil, nanotime(), [:], nil)setcpuprofilerate(int32())} else if cpuprof.on {setcpuprofilerate(0)cpuprof.on = falsecpuprof.addExtra()cpuprof.log.close()}unlock(&cpuprof.lock)}// add adds the stack trace to the profile.// It is called from signal handlers and other limited environments// and cannot allocate memory or acquire locks that might be// held at the time of the signal, nor can it use substantial amounts// of stack.////go:nowritebarrierrecfunc ( *cpuProfile) ( *unsafe.Pointer, []uintptr) {// Simple cas-lock to coordinate with setcpuprofilerate.for !prof.signalLock.CompareAndSwap(0, 1) {// TODO: Is it safe to osyield here? https://go.dev/issue/52672osyield()}if prof.hz.Load() != 0 { // implies cpuprof.log != nilif .numExtra > 0 || .lostExtra > 0 || .lostAtomic > 0 {.addExtra()}:= [1]uint64{1}// Note: write "knows" that the argument is &gp.labels,// because otherwise its write barrier behavior may not// be correct. See the long comment there before// changing the argument here.cpuprof.log.write(, nanotime(), [:], )}prof.signalLock.Store(0)}// addNonGo adds the non-Go stack trace to the profile.// It is called from a non-Go thread, so we cannot use much stack at all,// nor do anything that needs a g or an m.// In particular, we can't call cpuprof.log.write.// Instead, we copy the stack into cpuprof.extra,// which will be drained the next time a Go thread// gets the signal handling event.////go:nosplit//go:nowritebarrierrecfunc ( *cpuProfile) ( []uintptr) {// Simple cas-lock to coordinate with SetCPUProfileRate.// (Other calls to add or addNonGo should be blocked out// by the fact that only one SIGPROF can be handled by the// process at a time. If not, this lock will serialize those too.// The use of timer_create(2) on Linux to request process-targeted// signals may have changed this.)for !prof.signalLock.CompareAndSwap(0, 1) {// TODO: Is it safe to osyield here? https://go.dev/issue/52672osyield()}if cpuprof.numExtra+1+len() < len(cpuprof.extra) {:= cpuprof.numExtracpuprof.extra[] = uintptr(1 + len())copy(cpuprof.extra[+1:], )cpuprof.numExtra += 1 + len()} else {cpuprof.lostExtra++}prof.signalLock.Store(0)}// addExtra adds the "extra" profiling events,// queued by addNonGo, to the profile log.// addExtra is called either from a signal handler on a Go thread// or from an ordinary goroutine; either way it can use stack// and has a g. The world may be stopped, though.func ( *cpuProfile) () {// Copy accumulated non-Go profile events.:= [1]uint64{1}for := 0; < .numExtra; {.log.write(nil, 0, [:], .extra[+1:+int(.extra[])])+= int(.extra[])}.numExtra = 0// Report any lost events.if .lostExtra > 0 {:= [1]uint64{.lostExtra}:= [2]uintptr{abi.FuncPCABIInternal(_LostExternalCode) + sys.PCQuantum,abi.FuncPCABIInternal(_ExternalCode) + sys.PCQuantum,}.log.write(nil, 0, [:], [:]).lostExtra = 0}if .lostAtomic > 0 {:= [1]uint64{.lostAtomic}:= [2]uintptr{abi.FuncPCABIInternal(_LostSIGPROFDuringAtomic64) + sys.PCQuantum,abi.FuncPCABIInternal(_System) + sys.PCQuantum,}.log.write(nil, 0, [:], [:]).lostAtomic = 0}}// CPUProfile panics.// It formerly provided raw access to chunks of// a pprof-format profile generated by the runtime.// The details of generating that format have changed,// so this functionality has been removed.//// Deprecated: Use the runtime/pprof package,// or the handlers in the net/http/pprof package,// or the testing package's -test.cpuprofile flag instead.func () []byte {panic("CPUProfile no longer available")}//go:linkname runtime_pprof_runtime_cyclesPerSecond runtime/pprof.runtime_cyclesPerSecondfunc () int64 {return tickspersecond()}// readProfile, provided to runtime/pprof, returns the next chunk of// binary CPU profiling stack trace data, blocking until data is available.// If profiling is turned off and all the profile data accumulated while it was// on has been returned, readProfile returns eof=true.// The caller must save the returned data and tags before calling readProfile again.// The returned data contains a whole number of records, and tags contains// exactly one entry per record.////go:linkname runtime_pprof_readProfile runtime/pprof.readProfilefunc () ([]uint64, []unsafe.Pointer, bool) {lock(&cpuprof.lock):= cpuprof.logunlock(&cpuprof.lock):= profBufBlockingif GOOS == "darwin" || GOOS == "ios" {= profBufNonBlocking // For #61768; on Darwin notes are not async-signal-safe. See sigNoteSetup in os_darwin.go.}, , := .read()if len() == 0 && {lock(&cpuprof.lock)cpuprof.log = nilunlock(&cpuprof.lock)}return , ,}
![]() |
The pages are generated with Golds v0.6.7. (GOOS=linux GOARCH=amd64) Golds is a Go 101 project developed by Tapir Liu. PR and bug reports are welcome and can be submitted to the issue list. Please follow @Go100and1 (reachable from the left QR code) to get the latest news of Golds. |