// Copyright 2021 The Prometheus Authors // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. //go:build go1.17 // +build go1.17 package prometheus import ( "math" "reflect" "regexp" "runtime" "runtime/metrics" "sync" "testing" dto "github.com/prometheus/client_model/go" "github.com/prometheus/client_golang/prometheus/internal" ) func TestRmForMemStats(t *testing.T) { descs := bestEffortLookupRM(rmNamesForMemStatsMetrics) if got, want := len(descs), len(rmNamesForMemStatsMetrics); got != want { t.Errorf("got %d, want %d metrics", got, want) } for _, d := range descs { // We don't expect histograms there. if d.Kind == metrics.KindFloat64Histogram { t.Errorf("we don't expect to use histograms for MemStats metrics, got %v", d.Name) } } } func expectedBaseMetrics() map[string]struct{} { metrics := map[string]struct{}{} b := newBaseGoCollector() for _, m := range []string{ b.gcDesc.fqName, b.goInfoDesc.fqName, b.goroutinesDesc.fqName, b.gcLastTimeDesc.fqName, b.threadsDesc.fqName, } { metrics[m] = struct{}{} } return metrics } func addExpectedRuntimeMemStats(metrics map[string]struct{}) map[string]struct{} { for _, m := range goRuntimeMemStats() { metrics[m.desc.fqName] = struct{}{} } return metrics } func addExpectedRuntimeMetrics(metrics map[string]struct{}) map[string]struct{} { for _, m := range expectedRuntimeMetrics { metrics[m] = struct{}{} } return metrics } func TestGoCollector_ExposedMetrics(t *testing.T) { for _, tcase := range []struct { opts internal.GoCollectorOptions expectedFQNameSet map[string]struct{} }{ { opts: internal.GoCollectorOptions{ DisableMemStatsLikeMetrics: true, }, expectedFQNameSet: expectedBaseMetrics(), }, { // Default, only MemStats. expectedFQNameSet: addExpectedRuntimeMemStats(expectedBaseMetrics()), }, { // Get all runtime/metrics without MemStats. opts: internal.GoCollectorOptions{ DisableMemStatsLikeMetrics: true, RuntimeMetricRules: []internal.GoCollectorRule{ {Matcher: regexp.MustCompile("/.*")}, }, }, expectedFQNameSet: addExpectedRuntimeMetrics(expectedBaseMetrics()), }, { // Get all runtime/metrics and MemStats. opts: internal.GoCollectorOptions{ RuntimeMetricRules: []internal.GoCollectorRule{ {Matcher: regexp.MustCompile("/.*")}, }, }, expectedFQNameSet: addExpectedRuntimeMemStats(addExpectedRuntimeMetrics(expectedBaseMetrics())), }, } { if ok := t.Run("", func(t *testing.T) { goMetrics := collectGoMetrics(t, tcase.opts) goMetricSet := make(map[string]Metric) for _, m := range goMetrics { goMetricSet[m.Desc().fqName] = m } for i := range goMetrics { name := goMetrics[i].Desc().fqName if _, ok := tcase.expectedFQNameSet[name]; !ok { t.Errorf("found unpexpected metric %s", name) continue } } // Now iterate over the expected metrics and look for removals. for expectedName := range tcase.expectedFQNameSet { if _, ok := goMetricSet[expectedName]; !ok { t.Errorf("missing expected metric %s in collection", expectedName) continue } } }); !ok { return } } } var sink interface{} func TestBatchHistogram(t *testing.T) { goMetrics := collectGoMetrics(t, internal.GoCollectorOptions{ RuntimeMetricRules: []internal.GoCollectorRule{ {Matcher: regexp.MustCompile("/.*")}, }, }) var mhist Metric for _, m := range goMetrics { if m.Desc().fqName == "go_gc_heap_allocs_by_size_bytes" { mhist = m break } } if mhist == nil { t.Fatal("failed to find metric to test") } hist, ok := mhist.(*batchHistogram) if !ok { t.Fatal("found metric is not a runtime/metrics histogram") } // Make a bunch of allocations then do another collection. // // The runtime/metrics API tries to reuse memory where possible, // so make sure that we didn't hang on to any of that memory in // hist. countsCopy := make([]uint64, len(hist.counts)) copy(countsCopy, hist.counts) for i := 0; i < 100; i++ { sink = make([]byte, 128) } collectGoMetrics(t, defaultGoCollectorOptions()) for i, v := range hist.counts { if v != countsCopy[i] { t.Error("counts changed during new collection") break } } // Get the runtime/metrics copy. s := []metrics.Sample{ {Name: "/gc/heap/allocs-by-size:bytes"}, } metrics.Read(s) rmHist := s[0].Value.Float64Histogram() wantBuckets := internal.RuntimeMetricsBucketsForUnit(rmHist.Buckets, "bytes") // runtime/metrics histograms always have a +Inf bucket and are lower // bound inclusive. In contrast, we have an implicit +Inf bucket and // are upper bound inclusive, so we can chop off the first bucket // (since the conversion to upper bound inclusive will shift all buckets // down one index) and the +Inf for the last bucket. wantBuckets = wantBuckets[1 : len(wantBuckets)-1] // Check to make sure the output proto makes sense. pb := &dto.Metric{} hist.Write(pb) if math.IsInf(pb.Histogram.Bucket[len(pb.Histogram.Bucket)-1].GetUpperBound(), +1) { t.Errorf("found +Inf bucket") } if got := len(pb.Histogram.Bucket); got != len(wantBuckets) { t.Errorf("got %d buckets in protobuf, want %d", got, len(wantBuckets)) } for i, bucket := range pb.Histogram.Bucket { // runtime/metrics histograms are lower-bound inclusive, but we're // upper-bound inclusive. So just make sure the new inclusive upper // bound is somewhere close by (in some cases it's equal). wantBound := wantBuckets[i] if gotBound := *bucket.UpperBound; (wantBound-gotBound)/wantBound > 0.001 { t.Errorf("got bound %f, want within 0.1%% of %f", gotBound, wantBound) } // Make sure counts are cumulative. Because of the consistency guarantees // made by the runtime/metrics package, we're really not guaranteed to get // anything even remotely the same here. if i > 0 && *bucket.CumulativeCount < *pb.Histogram.Bucket[i-1].CumulativeCount { t.Error("cumulative counts are non-monotonic") } } } func collectGoMetrics(t *testing.T, opts internal.GoCollectorOptions) []Metric { t.Helper() c := NewGoCollector(func(o *internal.GoCollectorOptions) { o.DisableMemStatsLikeMetrics = opts.DisableMemStatsLikeMetrics o.RuntimeMetricSumForHist = opts.RuntimeMetricSumForHist o.RuntimeMetricRules = opts.RuntimeMetricRules }).(*goCollector) // Collect all metrics. ch := make(chan Metric) var wg sync.WaitGroup var metrics []Metric wg.Add(1) go func() { defer wg.Done() for metric := range ch { metrics = append(metrics, metric) } }() c.Collect(ch) close(ch) wg.Wait() return metrics } func TestMemStatsEquivalence(t *testing.T) { var msReal, msFake runtime.MemStats descs := bestEffortLookupRM(rmNamesForMemStatsMetrics) samples := make([]metrics.Sample, len(descs)) samplesMap := make(map[string]*metrics.Sample) for i := range descs { samples[i].Name = descs[i].Name samplesMap[descs[i].Name] = &samples[i] } // Force a GC cycle to try to reach a clean slate. runtime.GC() // Populate msReal. runtime.ReadMemStats(&msReal) // Populate msFake and hope that no GC happened in between (: metrics.Read(samples) memStatsFromRM(&msFake, samplesMap) // Iterate over them and make sure they're somewhat close. msRealValue := reflect.ValueOf(msReal) msFakeValue := reflect.ValueOf(msFake) typ := msRealValue.Type() for i := 0; i < msRealValue.NumField(); i++ { fr := msRealValue.Field(i) ff := msFakeValue.Field(i) if typ.Field(i).Name == "PauseTotalNs" || typ.Field(i).Name == "LastGC" { // We don't use those fields for metrics, // thus we are not interested in having this filled. continue } switch fr.Kind() { // Fields which we are interested in are all uint64s. // The only float64 field GCCPUFraction is by design omitted. case reflect.Uint64: vr := fr.Interface().(uint64) vf := ff.Interface().(uint64) if float64(vr-vf)/float64(vf) > 0.05 { t.Errorf("wrong value for %s: got %d, want %d", typ.Field(i).Name, vf, vr) } } } } func TestExpectedRuntimeMetrics(t *testing.T) { goMetrics := collectGoMetrics(t, internal.GoCollectorOptions{ DisableMemStatsLikeMetrics: true, RuntimeMetricRules: []internal.GoCollectorRule{ {Matcher: regexp.MustCompile("/.*")}, }, }) goMetricSet := make(map[string]Metric) for _, m := range goMetrics { goMetricSet[m.Desc().fqName] = m } descs := metrics.All() rmSet := make(map[string]struct{}) // Iterate over runtime-reported descriptions to find new metrics. for i := range descs { rmName := descs[i].Name rmSet[rmName] = struct{}{} // expectedRuntimeMetrics depends on Go version. expFQName, ok := expectedRuntimeMetrics[rmName] if !ok { t.Errorf("found new runtime/metrics metric %s", rmName) _, _, _, ok := internal.RuntimeMetricsToProm(&descs[i]) if !ok { t.Errorf("new metric has name that can't be converted, or has an unsupported Kind") } continue } _, ok = goMetricSet[expFQName] if !ok { t.Errorf("existing runtime/metrics metric %s (expected fq name %s) not collected", rmName, expFQName) continue } } // Now iterate over the expected metrics and look for removals. cardinality := 0 for rmName, fqName := range expectedRuntimeMetrics { if _, ok := rmSet[rmName]; !ok { t.Errorf("runtime/metrics metric %s removed", rmName) continue } if _, ok := goMetricSet[fqName]; !ok { t.Errorf("runtime/metrics metric %s not appearing under expected name %s", rmName, fqName) continue } // While we're at it, check to make sure expected cardinality lines // up, but at the point of the protobuf write to get as close to the // real deal as possible. // // Note that we filter out non-runtime/metrics metrics here, because // those are manually managed. var m dto.Metric if err := goMetricSet[fqName].Write(&m); err != nil { t.Errorf("writing metric %s: %v", fqName, err) continue } // N.B. These are the only fields populated by runtime/metrics metrics specifically. // Other fields are populated by e.g. GCStats metrics. switch { case m.Counter != nil: fallthrough case m.Gauge != nil: cardinality++ case m.Histogram != nil: cardinality += len(m.Histogram.Bucket) + 3 // + sum, count, and +inf default: t.Errorf("unexpected protobuf structure for metric %s", fqName) } } if t.Failed() { t.Log("a new Go version may have been detected, please run") t.Log("\tgo run gen_go_collector_metrics_set.go go1.X") t.Log("where X is the Go version you are currently using") } expectCardinality := expectedRuntimeMetricsCardinality if cardinality != expectCardinality { t.Errorf("unexpected cardinality for runtime/metrics metrics: got %d, want %d", cardinality, expectCardinality) } } func TestGoCollectorConcurrency(t *testing.T) { c := NewGoCollector().(*goCollector) // Set up multiple goroutines to Collect from the // same GoCollector. In race mode with GOMAXPROCS > 1, // this test should fail often if Collect is not // concurrent-safe. for i := 0; i < 4; i++ { go func() { ch := make(chan Metric) go func() { // Drain all metrics received until the // channel is closed. for range ch { } }() c.Collect(ch) close(ch) }() } }