From 0bb9a562509f6a9cba7dfef70db09976d069edaf Mon Sep 17 00:00:00 2001 From: Julius Volz Date: Wed, 31 Dec 2014 13:53:17 +0100 Subject: [PATCH] Remove extraction result type, simplify code. --- extraction/metricfamilyprocessor.go | 8 +- extraction/metricfamilyprocessor_test.go | 70 ++++---- extraction/processor.go | 38 +---- extraction/processor0_0_1.go | 20 +-- extraction/processor0_0_1_test.go | 198 +++++++++++------------ extraction/processor0_0_2.go | 19 +-- extraction/processor0_0_2_test.go | 197 +++++++++++----------- extraction/textprocessor_test.go | 62 ++++--- model/sample.go | 14 ++ 9 files changed, 283 insertions(+), 343 deletions(-) diff --git a/extraction/metricfamilyprocessor.go b/extraction/metricfamilyprocessor.go index 1be6432..af2bea6 100644 --- a/extraction/metricfamilyprocessor.go +++ b/extraction/metricfamilyprocessor.go @@ -101,7 +101,7 @@ func extractCounter(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error sample.Value = model.SampleValue(m.Counter.GetValue()) } - return out.Ingest(&Result{Samples: samples}) + return out.Ingest(samples) } func extractGauge(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error { @@ -132,7 +132,7 @@ func extractGauge(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error { sample.Value = model.SampleValue(m.Gauge.GetValue()) } - return out.Ingest(&Result{Samples: samples}) + return out.Ingest(samples) } func extractSummary(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error { @@ -194,7 +194,7 @@ func extractSummary(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error } } - return out.Ingest(&Result{Samples: samples}) + return out.Ingest(samples) } func extractUntyped(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error { @@ -225,5 +225,5 @@ func extractUntyped(out Ingester, o *ProcessOptions, f *dto.MetricFamily) error sample.Value = model.SampleValue(m.Untyped.GetValue()) } - return out.Ingest(&Result{Samples: samples}) + return out.Ingest(samples) } diff --git a/extraction/metricfamilyprocessor_test.go b/extraction/metricfamilyprocessor_test.go index 7bc6c7e..c4ffe90 100644 --- a/extraction/metricfamilyprocessor_test.go +++ b/extraction/metricfamilyprocessor_test.go @@ -25,11 +25,11 @@ var testTime = model.Now() type metricFamilyProcessorScenario struct { in string - expected, actual []*Result + expected, actual []model.Samples } -func (s *metricFamilyProcessorScenario) Ingest(r *Result) error { - s.actual = append(s.actual, r) +func (s *metricFamilyProcessorScenario) Ingest(samples model.Samples) error { + s.actual = append(s.actual, samples) return nil } @@ -50,10 +50,10 @@ func (s *metricFamilyProcessorScenario) test(t *testing.T, set int) { } for i, expected := range s.expected { - sort.Sort(s.actual[i].Samples) - sort.Sort(expected.Samples) + sort.Sort(s.actual[i]) + sort.Sort(expected) - if !expected.equal(s.actual[i]) { + if !expected.Equal(s.actual[i]) { t.Errorf("%d.%d. expected %s, got %s", set, i, expected, s.actual[i]) } } @@ -66,43 +66,39 @@ func TestMetricFamilyProcessor(t *testing.T) { }, { in: "\x8f\x01\n\rrequest_count\x12\x12Number of requests\x18\x00\"0\n#\n\x0fsome_label_name\x12\x10some_label_value\x1a\t\t\x00\x00\x00\x00\x00\x00E\xc0\"6\n)\n\x12another_label_name\x12\x13another_label_value\x1a\t\t\x00\x00\x00\x00\x00\x00U@", - expected: []*Result{ - { - Samples: model.Samples{ - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "request_count", "some_label_name": "some_label_value"}, - Value: -42, - Timestamp: testTime, - }, - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "request_count", "another_label_name": "another_label_value"}, - Value: 84, - Timestamp: testTime, - }, + expected: []model.Samples{ + model.Samples{ + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "request_count", "some_label_name": "some_label_value"}, + Value: -42, + Timestamp: testTime, + }, + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "request_count", "another_label_name": "another_label_value"}, + Value: 84, + Timestamp: testTime, }, }, }, }, { in: "\xb9\x01\n\rrequest_count\x12\x12Number of requests\x18\x02\"O\n#\n\x0fsome_label_name\x12\x10some_label_value\"(\x1a\x12\t\xaeG\xe1z\x14\xae\xef?\x11\x00\x00\x00\x00\x00\x00E\xc0\x1a\x12\t+\x87\x16\xd9\xce\xf7\xef?\x11\x00\x00\x00\x00\x00\x00U\xc0\"A\n)\n\x12another_label_name\x12\x13another_label_value\"\x14\x1a\x12\t\x00\x00\x00\x00\x00\x00\xe0?\x11\x00\x00\x00\x00\x00\x00$@", - expected: []*Result{ - { - Samples: model.Samples{ - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "request_count", "some_label_name": "some_label_value", "quantile": "0.99"}, - Value: -42, - Timestamp: testTime, - }, - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "request_count", "some_label_name": "some_label_value", "quantile": "0.999"}, - Value: -84, - Timestamp: testTime, - }, - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "request_count", "another_label_name": "another_label_value", "quantile": "0.5"}, - Value: 10, - Timestamp: testTime, - }, + expected: []model.Samples{ + model.Samples{ + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "request_count", "some_label_name": "some_label_value", "quantile": "0.99"}, + Value: -42, + Timestamp: testTime, + }, + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "request_count", "some_label_name": "some_label_value", "quantile": "0.999"}, + Value: -84, + Timestamp: testTime, + }, + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "request_count", "another_label_name": "another_label_value", "quantile": "0.5"}, + Value: 10, + Timestamp: testTime, }, }, }, diff --git a/extraction/processor.go b/extraction/processor.go index f7e0c29..89e8a6e 100644 --- a/extraction/processor.go +++ b/extraction/processor.go @@ -30,7 +30,7 @@ type ProcessOptions struct { // Ingester consumes result streams in whatever way is desired by the user. type Ingester interface { - Ingest(*Result) error + Ingest(model.Samples) error } // Processor is responsible for decoding the actual message responses from @@ -56,42 +56,6 @@ func labelSet(labels map[string]string) model.LabelSet { return labelset } -// Result encapsulates the outcome from processing samples from a source. -type Result struct { - Err error - Samples model.Samples -} - -func (r *Result) equal(o *Result) bool { - if r == o { - return true - } - - if r.Err != o.Err { - if r.Err == nil || o.Err == nil { - return false - } - - if r.Err.Error() != o.Err.Error() { - return false - } - } - - if len(r.Samples) != len(o.Samples) { - return false - } - - for i, mine := range r.Samples { - other := o.Samples[i] - - if !mine.Equal(other) { - return false - } - } - - return true -} - // A basic interface only useful in testing contexts for dispensing the time // in a controlled manner. type instantProvider interface { diff --git a/extraction/processor0_0_1.go b/extraction/processor0_0_1.go index 6c9760d..9f8bddc 100644 --- a/extraction/processor0_0_1.go +++ b/extraction/processor0_0_1.go @@ -77,11 +77,7 @@ func (p *processor001) ProcessSingle(in io.Reader, out Ingester, o *ProcessOptio case gauge001, counter001: sampleValue, ok := value.Value.(float64) if !ok { - err = fmt.Errorf("could not convert value from %s %s to float64", entity, value) - if err := out.Ingest(&Result{Err: err}); err != nil { - return err - } - continue + return fmt.Errorf("could not convert value from %s %s to float64", entity, value) } pendingSamples = append(pendingSamples, &model.Sample{ @@ -95,21 +91,13 @@ func (p *processor001) ProcessSingle(in io.Reader, out Ingester, o *ProcessOptio case histogram001: sampleValue, ok := value.Value.(map[string]interface{}) if !ok { - err = fmt.Errorf("could not convert value from %q to a map[string]interface{}", value.Value) - if err := out.Ingest(&Result{Err: err}); err != nil { - return err - } - continue + return fmt.Errorf("could not convert value from %q to a map[string]interface{}", value.Value) } for percentile, percentileValue := range sampleValue { individualValue, ok := percentileValue.(float64) if !ok { - err = fmt.Errorf("could not convert value from %q to a float64", percentileValue) - if err := out.Ingest(&Result{Err: err}); err != nil { - return err - } - continue + return fmt.Errorf("could not convert value from %q to a float64", percentileValue) } childMetric := make(map[model.LabelName]model.LabelValue, len(labels)+1) @@ -132,7 +120,7 @@ func (p *processor001) ProcessSingle(in io.Reader, out Ingester, o *ProcessOptio } } if len(pendingSamples) > 0 { - return out.Ingest(&Result{Samples: pendingSamples}) + return out.Ingest(pendingSamples) } return nil diff --git a/extraction/processor0_0_1_test.go b/extraction/processor0_0_1_test.go index 9ff530e..3ffaa04 100644 --- a/extraction/processor0_0_1_test.go +++ b/extraction/processor0_0_1_test.go @@ -29,12 +29,12 @@ var test001Time = model.Now() type testProcessor001ProcessScenario struct { in string - expected, actual []*Result + expected, actual []model.Samples err error } -func (s *testProcessor001ProcessScenario) Ingest(r *Result) error { - s.actual = append(s.actual, r) +func (s *testProcessor001ProcessScenario) Ingest(samples model.Samples) error { + s.actual = append(s.actual, samples) return nil } @@ -56,10 +56,10 @@ func (s *testProcessor001ProcessScenario) test(t testing.TB, set int) { } for i, expected := range s.expected { - sort.Sort(s.actual[i].Samples) - sort.Sort(expected.Samples) + sort.Sort(s.actual[i]) + sort.Sort(expected) - if !expected.equal(s.actual[i]) { + if !expected.Equal(s.actual[i]) { t.Errorf("%d.%d. expected %s, got %s", set, i, expected, s.actual[i]) } } @@ -73,101 +73,97 @@ func testProcessor001Process(t testing.TB) { }, { in: "test0_0_1-0_0_2.json", - expected: []*Result{ - { - Samples: model.Samples{ - &model.Sample{ - Metric: model.Metric{"service": "zed", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, - Value: 25, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"service": "bar", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, - Value: 25, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"service": "foo", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, - Value: 25, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 0.0459814091918713, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 78.48563317257356, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 15.890724674774395, - Timestamp: test001Time, - }, - &model.Sample{ - - Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 0.0459814091918713, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 78.48563317257356, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 15.890724674774395, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 0.6120456642749681, - Timestamp: test001Time, - }, - &model.Sample{ - - Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 97.31798360385088, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 84.63044031436561, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 1.355915069887731, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 109.89202084295582, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 160.21100853053224, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 1.772733213161236, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 109.99626121011262, - Timestamp: test001Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 172.49828748957728, - Timestamp: test001Time, - }, + expected: []model.Samples{ + model.Samples{ + &model.Sample{ + Metric: model.Metric{"service": "zed", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, + Value: 25, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"service": "bar", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, + Value: 25, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"service": "foo", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, + Value: 25, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 0.0459814091918713, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 78.48563317257356, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 15.890724674774395, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 0.0459814091918713, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 78.48563317257356, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 15.890724674774395, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 0.6120456642749681, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 97.31798360385088, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 84.63044031436561, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 1.355915069887731, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 109.89202084295582, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 160.21100853053224, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 1.772733213161236, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 109.99626121011262, + Timestamp: test001Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 172.49828748957728, + Timestamp: test001Time, }, }, }, diff --git a/extraction/processor0_0_2.go b/extraction/processor0_0_2.go index fbf660e..f1bb0b5 100644 --- a/extraction/processor0_0_2.go +++ b/extraction/processor0_0_2.go @@ -60,11 +60,7 @@ func (p *processor002) ProcessSingle(in io.Reader, out Ingester, o *ProcessOptio var values []counter002 if err := json.Unmarshal(entity.Metric.Values, &values); err != nil { - err := fmt.Errorf("could not extract %s value: %s", entity.Metric.Type, err) - if err := out.Ingest(&Result{Err: err}); err != nil { - return err - } - continue + return fmt.Errorf("could not extract %s value: %s", entity.Metric.Type, err) } for _, counter := range values { @@ -81,11 +77,7 @@ func (p *processor002) ProcessSingle(in io.Reader, out Ingester, o *ProcessOptio var values []histogram002 if err := json.Unmarshal(entity.Metric.Values, &values); err != nil { - err := fmt.Errorf("could not extract %s value: %s", entity.Metric.Type, err) - if err := out.Ingest(&Result{Err: err}); err != nil { - return err - } - continue + return fmt.Errorf("could not extract %s value: %s", entity.Metric.Type, err) } for _, histogram := range values { @@ -102,15 +94,12 @@ func (p *processor002) ProcessSingle(in io.Reader, out Ingester, o *ProcessOptio } default: - err := fmt.Errorf("unknown metric type %q", entity.Metric.Type) - if err := out.Ingest(&Result{Err: err}); err != nil { - return err - } + return fmt.Errorf("unknown metric type %q", entity.Metric.Type) } } if len(pendingSamples) > 0 { - return out.Ingest(&Result{Samples: pendingSamples}) + return out.Ingest(pendingSamples) } return nil diff --git a/extraction/processor0_0_2_test.go b/extraction/processor0_0_2_test.go index 4b7e592..6a7d458 100644 --- a/extraction/processor0_0_2_test.go +++ b/extraction/processor0_0_2_test.go @@ -30,12 +30,12 @@ var test002Time = model.Now() type testProcessor002ProcessScenario struct { in string - expected, actual []*Result + expected, actual []model.Samples err error } -func (s *testProcessor002ProcessScenario) Ingest(r *Result) error { - s.actual = append(s.actual, r) +func (s *testProcessor002ProcessScenario) Ingest(samples model.Samples) error { + s.actual = append(s.actual, samples) return nil } @@ -57,10 +57,10 @@ func (s *testProcessor002ProcessScenario) test(t testing.TB, set int) { } for i, expected := range s.expected { - sort.Sort(s.actual[i].Samples) - sort.Sort(expected.Samples) + sort.Sort(s.actual[i]) + sort.Sort(expected) - if !expected.equal(s.actual[i]) { + if !expected.Equal(s.actual[i]) { t.Fatalf("%d.%d. expected %s, got %s", set, i, expected, s.actual[i]) } } @@ -74,101 +74,98 @@ func testProcessor002Process(t testing.TB) { }, { in: "test0_0_1-0_0_2.json", - expected: []*Result{ - { - Samples: model.Samples{ - &model.Sample{ - Metric: model.Metric{"service": "zed", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, - Value: 25, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"service": "bar", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, - Value: 25, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"service": "foo", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, - Value: 25, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 0.0459814091918713, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 78.48563317257356, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 15.890724674774395, - Timestamp: test002Time, - }, - &model.Sample{ + expected: []model.Samples{ + model.Samples{ + &model.Sample{ + Metric: model.Metric{"service": "zed", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, + Value: 25, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"service": "bar", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, + Value: 25, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"service": "foo", model.MetricNameLabel: "rpc_calls_total", "job": "batch_job"}, + Value: 25, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 0.0459814091918713, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 78.48563317257356, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.010000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 15.890724674774395, + Timestamp: test002Time, + }, + &model.Sample{ - Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 0.0459814091918713, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 78.48563317257356, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 15.890724674774395, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 0.6120456642749681, - Timestamp: test002Time, - }, - &model.Sample{ - - Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 97.31798360385088, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 84.63044031436561, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 1.355915069887731, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 109.89202084295582, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 160.21100853053224, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, - Value: 1.772733213161236, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, - Value: 109.99626121011262, - Timestamp: test002Time, - }, - &model.Sample{ - Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, - Value: 172.49828748957728, - Timestamp: test002Time, - }, + Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 0.0459814091918713, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 78.48563317257356, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.050000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 15.890724674774395, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 0.6120456642749681, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 97.31798360385088, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.500000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 84.63044031436561, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 1.355915069887731, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 109.89202084295582, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.900000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 160.21100853053224, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "zed"}, + Value: 1.772733213161236, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "bar"}, + Value: 109.99626121011262, + Timestamp: test002Time, + }, + &model.Sample{ + Metric: model.Metric{"percentile": "0.990000", model.MetricNameLabel: "rpc_latency_microseconds", "service": "foo"}, + Value: 172.49828748957728, + Timestamp: test002Time, }, }, }, diff --git a/extraction/textprocessor_test.go b/extraction/textprocessor_test.go index e2c84ad..86e37fe 100644 --- a/extraction/textprocessor_test.go +++ b/extraction/textprocessor_test.go @@ -32,44 +32,40 @@ mf1{label="value1"} -3.14 123456 mf1{label="value2"} 42 mf2 4 ` - out = map[model.LabelValue]*Result{ - "mf1": { - Samples: model.Samples{ - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "mf1", "label": "value1"}, - Value: -3.14, - Timestamp: 123456, - }, - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "mf1", "label": "value2"}, - Value: 42, - Timestamp: ts, - }, + out = map[model.LabelValue]model.Samples{ + "mf1": model.Samples{ + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "mf1", "label": "value1"}, + Value: -3.14, + Timestamp: 123456, + }, + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "mf1", "label": "value2"}, + Value: 42, + Timestamp: ts, }, }, - "mf2": { - Samples: model.Samples{ - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "mf2"}, - Value: 3, - Timestamp: ts, - }, - &model.Sample{ - Metric: model.Metric{model.MetricNameLabel: "mf2"}, - Value: 4, - Timestamp: ts, - }, + "mf2": model.Samples{ + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "mf2"}, + Value: 3, + Timestamp: ts, + }, + &model.Sample{ + Metric: model.Metric{model.MetricNameLabel: "mf2"}, + Value: 4, + Timestamp: ts, }, }, } ) type testIngester struct { - results []*Result + results []model.Samples } -func (i *testIngester) Ingest(r *Result) error { - i.results = append(i.results, r) +func (i *testIngester) Ingest(s model.Samples) error { + i.results = append(i.results, s) return nil } @@ -88,16 +84,16 @@ func TestTextProcessor(t *testing.T) { t.Fatalf("Expected length %d, got %d", expected, got) } for _, r := range ingester.results { - expected, ok := out[r.Samples[0].Metric[model.MetricNameLabel]] + expected, ok := out[r[0].Metric[model.MetricNameLabel]] if !ok { t.Fatalf( "Unexpected metric name %q", - r.Samples[0].Metric[model.MetricNameLabel], + r[0].Metric[model.MetricNameLabel], ) } - sort.Sort(expected.Samples) - sort.Sort(r.Samples) - if !expected.equal(r) { + sort.Sort(expected) + sort.Sort(r) + if !expected.Equal(r) { t.Errorf("expected %s, got %s", expected, r) } } diff --git a/model/sample.go b/model/sample.go index 4b0eb07..735a70b 100644 --- a/model/sample.go +++ b/model/sample.go @@ -63,3 +63,17 @@ func (s Samples) Less(i, j int) bool { func (s Samples) Swap(i, j int) { s[i], s[j] = s[j], s[i] } + +// Equal compares two sets of samples and returns true if they are equal. +func (s Samples) Equal(o Samples) bool { + if len(s) != len(o) { + return false + } + + for i, sample := range s { + if !sample.Equal(o[i]) { + return false + } + } + return true +}