Prydin issue 8169 (#8357)

This commit is contained in:
Pontus Rydin 2020-11-03 13:09:37 -05:00 committed by GitHub
parent 1939e58b68
commit 89919631c5
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
1 changed files with 13 additions and 8 deletions

View File

@ -37,6 +37,8 @@ const maxSampleConst = 10 // Absolute maximum number of samples regardless of pe
const maxMetadataSamples = 100 // Number of resources to sample for metric metadata
const maxRealtimeMetrics = 50000 // Absolute maximum metrics per realtime query
const hwMarkTTL = time.Duration(4 * time.Hour)
type queryChunk []types.PerfQuerySpec
@ -322,13 +324,13 @@ func (e *Endpoint) reloadMetricNameMap(ctx context.Context) error {
return err
}
mn, err := client.CounterInfoByName(ctx)
mn, err := client.CounterInfoByKey(ctx)
if err != nil {
return err
}
e.metricNameLookup = make(map[int32]string)
for name, m := range mn {
e.metricNameLookup[m.Key] = name
for key, m := range mn {
e.metricNameLookup[key] = m.Name()
}
return nil
}
@ -889,6 +891,7 @@ func (e *Endpoint) chunkify(ctx context.Context, res *resourceKind, now time.Tim
}
pqs := make(queryChunk, 0, e.Parent.MaxQueryObjects)
numQs := 0
for _, object := range res.objects {
timeBuckets := make(map[int64]*types.PerfQuerySpec, 0)
@ -924,9 +927,9 @@ func (e *Endpoint) chunkify(ctx context.Context, res *resourceKind, now time.Tim
// Add this metric to the bucket
bucket.MetricId = append(bucket.MetricId, metric)
// Bucket filled to capacity? (Only applies to non real time)
// Bucket filled to capacity?
// OR if we're past the absolute maximum limit
if (!res.realTime && len(bucket.MetricId) >= maxMetrics) || len(bucket.MetricId) > 100000 {
if (!res.realTime && len(bucket.MetricId) >= maxMetrics) || len(bucket.MetricId) > maxRealtimeMetrics {
e.log.Debugf("Submitting partial query: %d metrics (%d remaining) of type %s for %s. Total objects %d",
len(bucket.MetricId), len(res.metrics)-metricIdx, res.name, e.URL.Host, len(res.objects))
@ -943,16 +946,18 @@ func (e *Endpoint) chunkify(ctx context.Context, res *resourceKind, now time.Tim
// Handle data in time bucket and submit job if we've reached the maximum number of object.
for _, bucket := range timeBuckets {
pqs = append(pqs, *bucket)
if (!res.realTime && len(pqs) > e.Parent.MaxQueryObjects) || len(pqs) > 100000 {
e.log.Debugf("Submitting final bucket job for %s: %d metrics", res.name, len(bucket.MetricId))
numQs += len(bucket.MetricId)
if (!res.realTime && numQs > e.Parent.MaxQueryObjects) || numQs > maxRealtimeMetrics {
e.log.Debugf("Submitting final bucket job for %s: %d metrics", res.name, numQs)
submitChunkJob(ctx, te, job, pqs)
pqs = make(queryChunk, 0, e.Parent.MaxQueryObjects)
numQs = 0
}
}
}
// Submit any jobs left in the queue
if len(pqs) > 0 {
e.log.Debugf("Submitting job for %s: %d objects", res.name, len(pqs))
e.log.Debugf("Submitting job for %s: %d objects, %d metrics", res.name, len(pqs), numQs)
submitChunkJob(ctx, te, job, pqs)
}