mirror of
https://github.com/k3s-io/kubernetes.git
synced 2025-07-23 11:50:44 +00:00
Use timing ratio histograms instead of sample-and-watermark histograms
This commit is contained in:
parent
8039f9d600
commit
0c0b7ca49f
@ -220,13 +220,6 @@ func TestApfRejectRequest(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestApfExemptRequest(t *testing.T) {
|
||||
epmetrics.Register()
|
||||
fcmetrics.Register()
|
||||
|
||||
// Wait for at least one sampling window to pass since creation of metrics.ReadWriteConcurrencyObserverPairGenerator,
|
||||
// so that an observation will cause some data to go into the Prometheus metrics.
|
||||
time.Sleep(time.Millisecond * 50)
|
||||
|
||||
server := newApfServerWithSingleRequest(t, decisionNoQueuingExecute)
|
||||
defer server.Close()
|
||||
|
||||
@ -240,19 +233,11 @@ func TestApfExemptRequest(t *testing.T) {
|
||||
|
||||
checkForExpectedMetrics(t, []string{
|
||||
"apiserver_current_inflight_requests",
|
||||
"apiserver_flowcontrol_read_vs_write_request_count_watermarks",
|
||||
"apiserver_flowcontrol_read_vs_write_request_count_samples",
|
||||
"apiserver_flowcontrol_read_vs_write_current_requests",
|
||||
})
|
||||
}
|
||||
|
||||
func TestApfExecuteRequest(t *testing.T) {
|
||||
epmetrics.Register()
|
||||
fcmetrics.Register()
|
||||
|
||||
// Wait for at least one sampling window to pass since creation of metrics.ReadWriteConcurrencyObserverPairGenerator,
|
||||
// so that an observation will cause some data to go into the Prometheus metrics.
|
||||
time.Sleep(time.Millisecond * 50)
|
||||
|
||||
server := newApfServerWithSingleRequest(t, decisionQueuingExecute)
|
||||
defer server.Close()
|
||||
|
||||
@ -267,19 +252,11 @@ func TestApfExecuteRequest(t *testing.T) {
|
||||
checkForExpectedMetrics(t, []string{
|
||||
"apiserver_current_inflight_requests",
|
||||
"apiserver_current_inqueue_requests",
|
||||
"apiserver_flowcontrol_read_vs_write_request_count_watermarks",
|
||||
"apiserver_flowcontrol_read_vs_write_request_count_samples",
|
||||
"apiserver_flowcontrol_read_vs_write_current_requests",
|
||||
})
|
||||
}
|
||||
|
||||
func TestApfExecuteMultipleRequests(t *testing.T) {
|
||||
epmetrics.Register()
|
||||
fcmetrics.Register()
|
||||
|
||||
// Wait for at least one sampling window to pass since creation of metrics.ReadWriteConcurrencyObserverPairGenerator,
|
||||
// so that an observation will cause some data to go into the Prometheus metrics.
|
||||
time.Sleep(time.Millisecond * 50)
|
||||
|
||||
concurrentRequests := 5
|
||||
preStartExecute, postStartExecute := &sync.WaitGroup{}, &sync.WaitGroup{}
|
||||
preEnqueue, postEnqueue := &sync.WaitGroup{}, &sync.WaitGroup{}
|
||||
@ -347,8 +324,7 @@ func TestApfExecuteMultipleRequests(t *testing.T) {
|
||||
checkForExpectedMetrics(t, []string{
|
||||
"apiserver_current_inflight_requests",
|
||||
"apiserver_current_inqueue_requests",
|
||||
"apiserver_flowcontrol_read_vs_write_request_count_watermarks",
|
||||
"apiserver_flowcontrol_read_vs_write_request_count_samples",
|
||||
"apiserver_flowcontrol_read_vs_write_current_requests",
|
||||
})
|
||||
}
|
||||
|
||||
|
@ -27,7 +27,6 @@ import (
|
||||
compbasemetrics "k8s.io/component-base/metrics"
|
||||
"k8s.io/component-base/metrics/legacyregistry"
|
||||
basemetricstestutil "k8s.io/component-base/metrics/testutil"
|
||||
"k8s.io/utils/clock"
|
||||
)
|
||||
|
||||
const (
|
||||
@ -40,7 +39,9 @@ const (
|
||||
priorityLevel = "priority_level"
|
||||
flowSchema = "flow_schema"
|
||||
phase = "phase"
|
||||
mark = "mark"
|
||||
LabelNamePhase = "phase"
|
||||
LabelValueWaiting = "waiting"
|
||||
LabelValueExecuting = "executing"
|
||||
)
|
||||
|
||||
var (
|
||||
@ -106,66 +107,45 @@ var (
|
||||
[]string{priorityLevel, flowSchema},
|
||||
)
|
||||
// PriorityLevelExecutionSeatsGaugeVec creates observers of seats occupied throughout execution for priority levels
|
||||
PriorityLevelExecutionSeatsGaugeVec = NewSampleAndWaterMarkHistogramsVec(clock.RealClock{}, time.Millisecond,
|
||||
&compbasemetrics.HistogramOpts{
|
||||
PriorityLevelExecutionSeatsGaugeVec = NewTimingRatioHistogramVec(
|
||||
&compbasemetrics.TimingHistogramOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: subsystem,
|
||||
Name: "priority_level_seat_count_samples",
|
||||
Help: "Periodic observations of number of seats occupied for any stage of execution (but only initial stage for WATCHes)",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1},
|
||||
Name: "priority_level_seat_utilization",
|
||||
Help: "Observations, at the end of every nanosecond, of utilization of seats for any stage of execution (but only initial stage for WATCHes)",
|
||||
// Buckets for both 0.99 and 1.0 mean PromQL's histogram_quantile will reveal saturation
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.95, 0.99, 1},
|
||||
ConstLabels: map[string]string{phase: "executing"},
|
||||
StabilityLevel: compbasemetrics.ALPHA,
|
||||
},
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: subsystem,
|
||||
Name: "priority_level_seat_count_watermarks",
|
||||
Help: "Watermarks of the number of seats occupied for any stage of execution (but only initial stage for WATCHes)",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1},
|
||||
ConstLabels: map[string]string{phase: "executing"},
|
||||
StabilityLevel: compbasemetrics.ALPHA,
|
||||
},
|
||||
[]string{priorityLevel},
|
||||
priorityLevel,
|
||||
)
|
||||
// PriorityLevelConcurrencyGaugeVec creates gauges of concurrency broken down by phase, priority level
|
||||
PriorityLevelConcurrencyGaugeVec = NewSampleAndWaterMarkHistogramsVec(clock.RealClock{}, time.Millisecond,
|
||||
&compbasemetrics.HistogramOpts{
|
||||
PriorityLevelConcurrencyGaugeVec = NewTimingRatioHistogramVec(
|
||||
&compbasemetrics.TimingHistogramOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: subsystem,
|
||||
Name: "priority_level_request_count_samples",
|
||||
Help: "Periodic observations of the number of requests waiting or in any stage of execution (but only initial stage for WATCHes)",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1},
|
||||
Name: "priority_level_request_utilization",
|
||||
Help: "Observations, at the end of every nanosecond, of number of requests (as a fraction of the relevant limit) waiting or in any stage of execution (but only initial stage for WATCHes)",
|
||||
// For executing: the denominator will be seats, so this metric will skew low.
|
||||
// FOr waiting: the denominiator is individual queue length limit, so this metric can go over 1. Issue #110160
|
||||
Buckets: []float64{0, 0.001, 0.0025, 0.005, 0.1, 0.25, 0.5, 0.75, 1, 10, 100},
|
||||
StabilityLevel: compbasemetrics.ALPHA,
|
||||
},
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: subsystem,
|
||||
Name: "priority_level_request_count_watermarks",
|
||||
Help: "Watermarks of the number of requests waiting or in any stage of execution (but only initial stage for WATCHes)",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1},
|
||||
StabilityLevel: compbasemetrics.ALPHA,
|
||||
},
|
||||
[]string{LabelNamePhase, priorityLevel},
|
||||
LabelNamePhase, priorityLevel,
|
||||
)
|
||||
// ReadWriteConcurrencyGaugeVec creates gauges of number of requests broken down by phase and mutating vs readonly
|
||||
ReadWriteConcurrencyGaugeVec = NewSampleAndWaterMarkHistogramsVec(clock.RealClock{}, time.Millisecond,
|
||||
&compbasemetrics.HistogramOpts{
|
||||
// ReadWriteConcurrencyPairVec creates gauges of number of requests broken down by phase and mutating vs readonly
|
||||
ReadWriteConcurrencyGaugeVec = NewTimingRatioHistogramVec(
|
||||
&compbasemetrics.TimingHistogramOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: subsystem,
|
||||
Name: "read_vs_write_request_count_samples",
|
||||
Help: "Periodic observations of the number of requests waiting or in regular stage of execution",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1},
|
||||
Name: "read_vs_write_current_requests",
|
||||
Help: "Observations, at the end of every nanosecond, of the number of requests (as a fraction of the relevant limit, if max-in-flight filter is being used) waiting or in regular stage of execution",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 0.95, 0.99, 1, 3, 10, 30, 100, 300, 1000, 3000},
|
||||
// TODO: something about the utilization vs count irregularity. Issue #109846
|
||||
StabilityLevel: compbasemetrics.ALPHA,
|
||||
},
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: namespace,
|
||||
Subsystem: subsystem,
|
||||
Name: "read_vs_write_request_count_watermarks",
|
||||
Help: "Watermarks of the number of requests waiting or in regular stage of execution",
|
||||
Buckets: []float64{0, 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1},
|
||||
StabilityLevel: compbasemetrics.ALPHA,
|
||||
},
|
||||
[]string{LabelNamePhase, requestKind},
|
||||
LabelNamePhase, requestKind,
|
||||
)
|
||||
apiserverCurrentR = compbasemetrics.NewGaugeVec(
|
||||
&compbasemetrics.GaugeOpts{
|
||||
|
@ -1,204 +0,0 @@
|
||||
/*
|
||||
Copyright 2019 The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package metrics
|
||||
|
||||
import (
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
compbasemetrics "k8s.io/component-base/metrics"
|
||||
"k8s.io/klog/v2"
|
||||
"k8s.io/utils/clock"
|
||||
)
|
||||
|
||||
const (
|
||||
labelNameMark = "mark"
|
||||
labelValueLo = "low"
|
||||
labelValueHi = "high"
|
||||
LabelNamePhase = "phase"
|
||||
LabelValueWaiting = "waiting"
|
||||
LabelValueExecuting = "executing"
|
||||
)
|
||||
|
||||
// SampleAndWaterMarkObserverVec creates RatioedGauges that
|
||||
// populate histograms of samples and low- and high-water-marks. The
|
||||
// generator has a samplePeriod, and the histograms get an observation
|
||||
// every samplePeriod. The sampling windows are quantized based on
|
||||
// the monotonic rather than wall-clock times. The `t0` field is
|
||||
// there so to provide a baseline for monotonic clock differences.
|
||||
type SampleAndWaterMarkObserverVec struct {
|
||||
*sampleAndWaterMarkObserverVec
|
||||
}
|
||||
|
||||
type sampleAndWaterMarkObserverVec struct {
|
||||
clock clock.PassiveClock
|
||||
t0 time.Time
|
||||
samplePeriod time.Duration
|
||||
samples *compbasemetrics.HistogramVec
|
||||
waterMarks *compbasemetrics.HistogramVec
|
||||
}
|
||||
|
||||
var _ RatioedGaugeVec = SampleAndWaterMarkObserverVec{}
|
||||
|
||||
// NewSampleAndWaterMarkHistogramsVec makes a new one
|
||||
func NewSampleAndWaterMarkHistogramsVec(clock clock.PassiveClock, samplePeriod time.Duration, sampleOpts, waterMarkOpts *compbasemetrics.HistogramOpts, labelNames []string) SampleAndWaterMarkObserverVec {
|
||||
return SampleAndWaterMarkObserverVec{
|
||||
&sampleAndWaterMarkObserverVec{
|
||||
clock: clock,
|
||||
t0: clock.Now(),
|
||||
samplePeriod: samplePeriod,
|
||||
samples: compbasemetrics.NewHistogramVec(sampleOpts, labelNames),
|
||||
waterMarks: compbasemetrics.NewHistogramVec(waterMarkOpts, append([]string{labelNameMark}, labelNames...)),
|
||||
}}
|
||||
}
|
||||
|
||||
func (swg *sampleAndWaterMarkObserverVec) quantize(when time.Time) int64 {
|
||||
return int64(when.Sub(swg.t0) / swg.samplePeriod)
|
||||
}
|
||||
|
||||
// NewForLabelValuesSafe makes a new RatioedGauge
|
||||
func (swg *sampleAndWaterMarkObserverVec) NewForLabelValuesSafe(initialNumerator, initialDenominator float64, labelValues []string) RatioedGauge {
|
||||
ratio := initialNumerator / initialDenominator
|
||||
when := swg.clock.Now()
|
||||
return &sampleAndWaterMarkHistograms{
|
||||
sampleAndWaterMarkObserverVec: swg,
|
||||
labelValues: labelValues,
|
||||
loLabelValues: append([]string{labelValueLo}, labelValues...),
|
||||
hiLabelValues: append([]string{labelValueHi}, labelValues...),
|
||||
denominator: initialDenominator,
|
||||
sampleAndWaterMarkAccumulator: sampleAndWaterMarkAccumulator{
|
||||
lastSet: when,
|
||||
lastSetInt: swg.quantize(when),
|
||||
numerator: initialNumerator,
|
||||
ratio: ratio,
|
||||
loRatio: ratio,
|
||||
hiRatio: ratio,
|
||||
}}
|
||||
}
|
||||
|
||||
func (swg *sampleAndWaterMarkObserverVec) metrics() Registerables {
|
||||
return Registerables{swg.samples, swg.waterMarks}
|
||||
}
|
||||
|
||||
type sampleAndWaterMarkHistograms struct {
|
||||
*sampleAndWaterMarkObserverVec
|
||||
labelValues []string
|
||||
loLabelValues, hiLabelValues []string
|
||||
|
||||
sync.Mutex
|
||||
denominator float64
|
||||
sampleAndWaterMarkAccumulator
|
||||
}
|
||||
|
||||
type sampleAndWaterMarkAccumulator struct {
|
||||
lastSet time.Time
|
||||
lastSetInt int64 // lastSet / samplePeriod
|
||||
numerator float64
|
||||
ratio float64 // numerator/denominator
|
||||
loRatio, hiRatio float64
|
||||
}
|
||||
|
||||
var _ RatioedGauge = (*sampleAndWaterMarkHistograms)(nil)
|
||||
|
||||
func (saw *sampleAndWaterMarkHistograms) Set(numerator float64) {
|
||||
saw.innerSet(func() {
|
||||
saw.numerator = numerator
|
||||
})
|
||||
}
|
||||
|
||||
func (saw *sampleAndWaterMarkHistograms) Add(deltaNumerator float64) {
|
||||
saw.innerSet(func() {
|
||||
saw.numerator += deltaNumerator
|
||||
})
|
||||
}
|
||||
func (saw *sampleAndWaterMarkHistograms) Sub(deltaNumerator float64) {
|
||||
saw.innerSet(func() {
|
||||
saw.numerator -= deltaNumerator
|
||||
})
|
||||
}
|
||||
|
||||
func (saw *sampleAndWaterMarkHistograms) Inc() {
|
||||
saw.innerSet(func() {
|
||||
saw.numerator += 1
|
||||
})
|
||||
}
|
||||
func (saw *sampleAndWaterMarkHistograms) Dec() {
|
||||
saw.innerSet(func() {
|
||||
saw.numerator -= 1
|
||||
})
|
||||
}
|
||||
|
||||
func (saw *sampleAndWaterMarkHistograms) SetToCurrentTime() {
|
||||
saw.innerSet(func() {
|
||||
saw.numerator = float64(saw.clock.Now().Sub(time.Unix(0, 0)))
|
||||
})
|
||||
}
|
||||
|
||||
func (saw *sampleAndWaterMarkHistograms) SetDenominator(denominator float64) {
|
||||
saw.innerSet(func() {
|
||||
saw.denominator = denominator
|
||||
})
|
||||
}
|
||||
|
||||
func (saw *sampleAndWaterMarkHistograms) innerSet(updateNumeratorOrDenominator func()) {
|
||||
when, whenInt, acc, wellOrdered := func() (time.Time, int64, sampleAndWaterMarkAccumulator, bool) {
|
||||
saw.Lock()
|
||||
defer saw.Unlock()
|
||||
// Moved these variables here to tiptoe around https://github.com/golang/go/issues/43570 for #97685
|
||||
when := saw.clock.Now()
|
||||
whenInt := saw.quantize(when)
|
||||
acc := saw.sampleAndWaterMarkAccumulator
|
||||
wellOrdered := !when.Before(acc.lastSet)
|
||||
updateNumeratorOrDenominator()
|
||||
saw.ratio = saw.numerator / saw.denominator
|
||||
if wellOrdered {
|
||||
if acc.lastSetInt < whenInt {
|
||||
saw.loRatio, saw.hiRatio = acc.ratio, acc.ratio
|
||||
saw.lastSetInt = whenInt
|
||||
}
|
||||
saw.lastSet = when
|
||||
}
|
||||
// `wellOrdered` should always be true because we are using
|
||||
// monotonic clock readings and they never go backwards. Yet
|
||||
// very small backwards steps (under 1 microsecond) have been
|
||||
// observed
|
||||
// (https://github.com/kubernetes/kubernetes/issues/96459).
|
||||
// In the backwards case, treat the current reading as if it
|
||||
// had occurred at time `saw.lastSet` and log an error. It
|
||||
// would be wrong to update `saw.lastSet` in this case because
|
||||
// that plants a time bomb for future updates to
|
||||
// `saw.lastSetInt`.
|
||||
if saw.ratio < saw.loRatio {
|
||||
saw.loRatio = saw.ratio
|
||||
} else if saw.ratio > saw.hiRatio {
|
||||
saw.hiRatio = saw.ratio
|
||||
}
|
||||
return when, whenInt, acc, wellOrdered
|
||||
}()
|
||||
if !wellOrdered {
|
||||
lastSetS := acc.lastSet.String()
|
||||
whenS := when.String()
|
||||
klog.Errorf("Time went backwards from %s to %s for labelValues=%#+v", lastSetS, whenS, saw.labelValues)
|
||||
}
|
||||
for acc.lastSetInt < whenInt {
|
||||
saw.samples.WithLabelValues(saw.labelValues...).Observe(acc.ratio)
|
||||
saw.waterMarks.WithLabelValues(saw.loLabelValues...).Observe(acc.loRatio)
|
||||
saw.waterMarks.WithLabelValues(saw.hiLabelValues...).Observe(acc.hiRatio)
|
||||
acc.lastSetInt++
|
||||
acc.loRatio, acc.hiRatio = acc.ratio, acc.ratio
|
||||
}
|
||||
}
|
@ -1,190 +0,0 @@
|
||||
/*
|
||||
Copyright 2021 The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package metrics
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"math/rand"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
compbasemetrics "k8s.io/component-base/metrics"
|
||||
"k8s.io/klog/v2"
|
||||
testclock "k8s.io/utils/clock/testing"
|
||||
)
|
||||
|
||||
const (
|
||||
samplesHistName = "sawtestsamples"
|
||||
samplingPeriod = time.Millisecond
|
||||
ddtRangeCentiPeriods = 300
|
||||
ddtOffsetCentiPeriods = 50
|
||||
numIterations = 100
|
||||
)
|
||||
|
||||
var errMetricNotFound = errors.New("not found")
|
||||
|
||||
/* TestSampler does a rough behavioral test of the sampling in a
|
||||
SampleAndWatermarkHistograms. The test creates one and exercises
|
||||
it, checking that the count in the sampling histogram is correct at
|
||||
each step. The sampling histogram is expected to get one
|
||||
observation at the end of each sampling period. A fake clock is
|
||||
used, and the exercise consists of repeatedly changing that fake
|
||||
clock by an amount of time chosen uniformly at random from a range
|
||||
that goes from a little negative to somewhat more than two sampling
|
||||
periods. The negative changes are included because small negative
|
||||
changes have been observed in real monotonic clock readings (see
|
||||
issue #96459) and we want to test that they are properly tolerated.
|
||||
The designed toleration is to pretend that the clock did not
|
||||
change, until it resumes net forward progress.
|
||||
*/
|
||||
func TestSampler(t *testing.T) {
|
||||
t0 := time.Now()
|
||||
clk := testclock.NewFakePassiveClock(t0)
|
||||
buckets := []float64{0, 1}
|
||||
gen := NewSampleAndWaterMarkHistogramsVec(clk, samplingPeriod,
|
||||
&compbasemetrics.HistogramOpts{Name: samplesHistName, Buckets: buckets},
|
||||
&compbasemetrics.HistogramOpts{Name: "marks", Buckets: buckets},
|
||||
[]string{})
|
||||
saw := gen.NewForLabelValuesSafe(0, 1, []string{})
|
||||
toRegister := gen.metrics()
|
||||
registry := compbasemetrics.NewKubeRegistry()
|
||||
for _, reg := range toRegister {
|
||||
registry.MustRegister(reg)
|
||||
}
|
||||
// `dt` is the admitted cumulative difference in fake time
|
||||
// since the start of the test. "admitted" means this is
|
||||
// never allowed to decrease, which matches the designed
|
||||
// toleration for negative monotonic clock changes.
|
||||
var dt time.Duration
|
||||
// `t1` is the current fake time
|
||||
t1 := t0.Add(dt)
|
||||
klog.Infof("Expect about %v warnings about time going backwards; this is fake time deliberately misbehaving.", (numIterations*ddtOffsetCentiPeriods)/ddtRangeCentiPeriods)
|
||||
t.Logf("t0=%s", t0)
|
||||
for i := 0; i < numIterations; i++ {
|
||||
// `ddt` is the next step to take in fake time
|
||||
ddt := time.Duration(rand.Intn(ddtRangeCentiPeriods)-ddtOffsetCentiPeriods) * samplingPeriod / 100
|
||||
t1 = t1.Add(ddt)
|
||||
diff := t1.Sub(t0)
|
||||
if diff > dt {
|
||||
dt = diff
|
||||
}
|
||||
clk.SetTime(t1)
|
||||
saw.Set(1)
|
||||
expectedCount := int64(dt / samplingPeriod)
|
||||
actualCount, err := getHistogramCount(registry, samplesHistName)
|
||||
if err != nil && !(err == errMetricNotFound && expectedCount == 0) {
|
||||
t.Fatalf("For t0=%s, t1=%s, failed to getHistogramCount: %#+v", t0, t1, err)
|
||||
}
|
||||
t.Logf("For i=%d, ddt=%s, t1=%s, diff=%s, dt=%s, count=%d", i, ddt, t1, diff, dt, actualCount)
|
||||
if expectedCount != actualCount {
|
||||
t.Errorf("For i=%d, t0=%s, ddt=%s, t1=%s, expectedCount=%d, actualCount=%d", i, t0, ddt, t1, expectedCount, actualCount)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/* getHistogramCount returns the count of the named histogram or an error (if any) */
|
||||
func getHistogramCount(registry compbasemetrics.KubeRegistry, metricName string) (int64, error) {
|
||||
mfs, err := registry.Gather()
|
||||
if err != nil {
|
||||
return 0, fmt.Errorf("failed to gather metrics: %w", err)
|
||||
}
|
||||
for _, mf := range mfs {
|
||||
thisName := mf.GetName()
|
||||
if thisName != metricName {
|
||||
continue
|
||||
}
|
||||
metric := mf.GetMetric()[0]
|
||||
hist := metric.GetHistogram()
|
||||
if hist == nil {
|
||||
return 0, errors.New("dto.Metric has nil Histogram")
|
||||
}
|
||||
if hist.SampleCount == nil {
|
||||
return 0, errors.New("dto.Histogram has nil SampleCount")
|
||||
}
|
||||
return int64(*hist.SampleCount), nil
|
||||
}
|
||||
return 0, errMetricNotFound
|
||||
}
|
||||
|
||||
func BenchmarkSampleAndWaterMarkHistogramsVecEltSafeEarly(b *testing.B) {
|
||||
b.StopTimer()
|
||||
now := time.Now()
|
||||
clk := testclock.NewFakePassiveClock(now)
|
||||
thv := NewSampleAndWaterMarkHistogramsVec(clk, time.Millisecond,
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: "testns",
|
||||
Subsystem: "testsubsys",
|
||||
Name: "samplehist",
|
||||
Help: "Me",
|
||||
Buckets: []float64{1, 2, 4, 8, 16, 32},
|
||||
},
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: "testns",
|
||||
Subsystem: "testsubsys",
|
||||
Name: "markhist",
|
||||
Help: "Me",
|
||||
Buckets: []float64{1, 2, 4, 8, 16, 32},
|
||||
},
|
||||
[]string{"labelname"})
|
||||
th := thv.NewForLabelValuesSafe(0, 3, []string{"labelvalue"})
|
||||
registry := compbasemetrics.NewKubeRegistry()
|
||||
registry.MustRegister(thv.metrics()...)
|
||||
var x int
|
||||
b.StartTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
delta := (i % 6) + 1
|
||||
now = now.Add(time.Duration(delta) * time.Millisecond)
|
||||
clk.SetTime(now)
|
||||
th.Set(float64(x))
|
||||
x = (x + i) % 60
|
||||
}
|
||||
}
|
||||
|
||||
func BenchmarkSampleAndWaterMarkHistogramsVecEltSafeLate(b *testing.B) {
|
||||
b.StopTimer()
|
||||
now := time.Now()
|
||||
clk := testclock.NewFakePassiveClock(now)
|
||||
thv := NewSampleAndWaterMarkHistogramsVec(clk, time.Millisecond,
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: "testns",
|
||||
Subsystem: "testsubsys",
|
||||
Name: "samplehist",
|
||||
Help: "Me",
|
||||
Buckets: []float64{1, 2, 4, 8, 16, 32},
|
||||
},
|
||||
&compbasemetrics.HistogramOpts{
|
||||
Namespace: "testns",
|
||||
Subsystem: "testsubsys",
|
||||
Name: "markhist",
|
||||
Help: "Me",
|
||||
Buckets: []float64{1, 2, 4, 8, 16, 32},
|
||||
},
|
||||
[]string{"labelname"})
|
||||
registry := compbasemetrics.NewKubeRegistry()
|
||||
registry.MustRegister(thv.metrics()...)
|
||||
th := thv.NewForLabelValuesSafe(0, 3, []string{"labelvalue"})
|
||||
var x int
|
||||
b.StartTimer()
|
||||
for i := 0; i < b.N; i++ {
|
||||
delta := (i % 6) + 1
|
||||
now = now.Add(time.Duration(delta) * time.Millisecond)
|
||||
clk.SetTime(now)
|
||||
th.Set(float64(x))
|
||||
x = (x + i) % 60
|
||||
}
|
||||
}
|
@ -17,6 +17,8 @@ limitations under the License.
|
||||
package metrics
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"math/rand"
|
||||
"testing"
|
||||
"time"
|
||||
@ -26,6 +28,14 @@ import (
|
||||
testclock "k8s.io/utils/clock/testing"
|
||||
)
|
||||
|
||||
const (
|
||||
ddtRangeCentiPeriods = 300
|
||||
ddtOffsetCentiPeriods = 50
|
||||
numIterations = 100
|
||||
)
|
||||
|
||||
var errMetricNotFound = errors.New("not found")
|
||||
|
||||
func TestTimingRatioHistogramVecElementSimple(t *testing.T) {
|
||||
testHistogramName := "vec_element_simple"
|
||||
t0 := time.Now()
|
||||
@ -130,6 +140,30 @@ func exerciseTimingRatioHistogram(t *testing.T, histogramName string, t0 time.Ti
|
||||
}
|
||||
}
|
||||
|
||||
/* getHistogramCount returns the count of the named histogram or an error (if any) */
|
||||
func getHistogramCount(registry compbasemetrics.KubeRegistry, metricName string) (int64, error) {
|
||||
mfs, err := registry.Gather()
|
||||
if err != nil {
|
||||
return 0, fmt.Errorf("failed to gather metrics: %w", err)
|
||||
}
|
||||
for _, mf := range mfs {
|
||||
thisName := mf.GetName()
|
||||
if thisName != metricName {
|
||||
continue
|
||||
}
|
||||
metric := mf.GetMetric()[0]
|
||||
hist := metric.GetHistogram()
|
||||
if hist == nil {
|
||||
return 0, errors.New("dto.Metric has nil Histogram")
|
||||
}
|
||||
if hist.SampleCount == nil {
|
||||
return 0, errors.New("dto.Histogram has nil SampleCount")
|
||||
}
|
||||
return int64(*hist.SampleCount), nil
|
||||
}
|
||||
return 0, errMetricNotFound
|
||||
}
|
||||
|
||||
func BenchmarkTimingRatioHistogram(b *testing.B) {
|
||||
b.StopTimer()
|
||||
now := time.Now()
|
||||
|
Loading…
Reference in New Issue
Block a user