diff --git a/.golangci.yml b/.golangci.yml index 193113d39..ab110bfd4 100644 --- a/.golangci.yml +++ b/.golangci.yml @@ -152,6 +152,19 @@ linters-settings: # Maximum allowed permissions mode for os.WriteFile and ioutil.WriteFile # Default: "0600" G306: "0640" + govet: + settings: + ## Check the logging function like it would be a printf + printf: + funcs: + - (github.com/influxdata/telegraf.Logger).Debugf + - (github.com/influxdata/telegraf.Logger).Infof + - (github.com/influxdata/telegraf.Logger).Warnf + - (github.com/influxdata/telegraf.Logger).Errorf + - (github.com/influxdata/telegraf.Logger).Debug + - (github.com/influxdata/telegraf.Logger).Info + - (github.com/influxdata/telegraf.Logger).Warn + - (github.com/influxdata/telegraf.Logger).Error lll: # Max line length, lines longer will be reported. # '\t' is counted as 1 character by default, and can be changed with the tab-width option. diff --git a/plugins/common/opcua/client.go b/plugins/common/opcua/client.go index 412d108c9..03229c303 100644 --- a/plugins/common/opcua/client.go +++ b/plugins/common/opcua/client.go @@ -169,7 +169,7 @@ func (o *OpcUAClient) Connect() error { } if o.Client != nil { - o.Log.Warnf("Closing connection due to Connect called while already instantiated", u) + o.Log.Warnf("Closing connection to %q as already connected", u) if err := o.Client.Close(); err != nil { // Only log the error but to not bail-out here as this prevents // reconnections for multiple parties (see e.g. #9523). diff --git a/plugins/inputs/ctrlx_datalayer/ctrlx_datalayer.go b/plugins/inputs/ctrlx_datalayer/ctrlx_datalayer.go index 08e4107df..eaca80a08 100644 --- a/plugins/inputs/ctrlx_datalayer/ctrlx_datalayer.go +++ b/plugins/inputs/ctrlx_datalayer/ctrlx_datalayer.go @@ -147,7 +147,7 @@ func (c *CtrlXDataLayer) addMetric(se *sseclient.SseEvent, sub *Subscription) { c.Log.Debug("Received keepalive event") default: // Received a yet unsupported event type - c.Log.Debug("Received unsupported event: %q", se.Event) + c.Log.Debugf("Received unsupported event: %q", se.Event) } } diff --git a/plugins/inputs/gnmi/handler.go b/plugins/inputs/gnmi/handler.go index e835c7cc4..039a7790c 100644 --- a/plugins/inputs/gnmi/handler.go +++ b/plugins/inputs/gnmi/handler.go @@ -98,10 +98,10 @@ func (h *handler) subscribeGNMI(ctx context.Context, acc telegraf.Accumulator, t if h.trace { buf, err := protojson.Marshal(reply) if err != nil { - h.log.Debugf("marshal failed: %v", err) + h.log.Debugf("Marshal failed: %v", err) } else { t := reply.GetUpdate().GetTimestamp() - h.log.Debugf("update_%v: %s", t, string(buf)) + h.log.Debugf("Got update_%v: %s", t, string(buf)) } } if response, ok := reply.Response.(*gnmiLib.SubscribeResponse_Update); ok { @@ -137,7 +137,7 @@ func (h *handler) handleSubscribeResponseUpdate(acc telegraf.Accumulator, respon // unmarshal extention err := proto.Unmarshal(currentExt, juniperHeader) if err != nil { - h.log.Errorf("unmarshal gnmi Juniper Header extention failed: %w", err) + h.log.Errorf("unmarshal gnmi Juniper Header extension failed: %v", err) break } // Add only relevant Tags from the Juniper Header extention. @@ -156,7 +156,7 @@ func (h *handler) handleSubscribeResponseUpdate(acc telegraf.Accumulator, respon var origin string var err error if origin, prefix, prefixAliasPath, err = handlePath(response.Update.Prefix, prefixTags, h.aliases, ""); err != nil { - h.log.Errorf("handling path %q failed: %v", response.Update.Prefix, err) + h.log.Errorf("Handling path %q failed: %v", response.Update.Prefix, err) } prefix = origin + prefix } @@ -187,7 +187,7 @@ func (h *handler) handleSubscribeResponseUpdate(acc telegraf.Accumulator, respon } h.log.Debugf("Tag-subscription update for %q: %+v", tagSub.Name, update) if err := h.tagStore.insert(tagSub, fullPath, fields, tags); err != nil { - h.log.Errorf("inserting tag failed: %w", err) + h.log.Errorf("Inserting tag failed: %v", err) } tagUpdate = true break @@ -259,7 +259,7 @@ func (h *handler) handleSubscribeResponseUpdate(acc telegraf.Accumulator, respon key = strings.TrimLeft(key, "/.") } if key == "" { - h.log.Errorf("invalid empty path: %q", k) + h.log.Errorf("Invalid empty path: %q", k) continue } grouper.Add(name, tags, timestamp, key, v) @@ -276,11 +276,11 @@ func (h *handler) handleSubscribeResponseUpdate(acc telegraf.Accumulator, respon func (h *handler) handleTelemetryField(update *gnmiLib.Update, tags map[string]string, prefix string) (string, map[string]interface{}) { _, gpath, aliasPath, err := handlePath(update.Path, tags, h.aliases, prefix) if err != nil { - h.log.Errorf("handling path %q failed: %v", update.Path, err) + h.log.Errorf("Handling path %q failed: %v", update.Path, err) } fields, err := gnmiToFields(strings.Replace(gpath, "-", "_", -1), update.Val) if err != nil { - h.log.Errorf("error parsing update value %q: %v", update.Val, err) + h.log.Errorf("Error parsing update value %q: %v", update.Val, err) } return aliasPath, fields } diff --git a/plugins/inputs/google_cloud_storage/google_cloud_storage.go b/plugins/inputs/google_cloud_storage/google_cloud_storage.go index 90034b58f..11e30e58e 100644 --- a/plugins/inputs/google_cloud_storage/google_cloud_storage.go +++ b/plugins/inputs/google_cloud_storage/google_cloud_storage.go @@ -88,7 +88,7 @@ func (gcs *GCS) Gather(acc telegraf.Accumulator) error { } if err != nil { - gcs.Log.Errorf("Error during iteration of keys", err) + gcs.Log.Errorf("Error during iteration of keys: %v", err) return err } @@ -279,6 +279,6 @@ func init() { func (gcs *GCS) closeReader(r *storage.Reader) { if err := r.Close(); err != nil { - gcs.Log.Errorf("Could not close reader", err) + gcs.Log.Errorf("Could not close reader: %v", err) } } diff --git a/plugins/inputs/intel_pmu/intel_pmu.go b/plugins/inputs/intel_pmu/intel_pmu.go index ab4a52a39..14f66d1fd 100644 --- a/plugins/inputs/intel_pmu/intel_pmu.go +++ b/plugins/inputs/intel_pmu/intel_pmu.go @@ -190,7 +190,7 @@ func (i *IntelPMU) checkFileDescriptors() error { // maximum file descriptors enforced on a kernel level maxFd, err := readMaxFD(i.fileInfo) if err != nil { - i.Log.Warnf("cannot obtain number of available file descriptors: %v", err) + i.Log.Warnf("Cannot obtain number of available file descriptors: %v", err) } else if allFd > maxFd { return fmt.Errorf("required file descriptors number `%d` exceeds maximum number of available file descriptors `%d`"+ ": consider increasing the maximum number", allFd, maxFd) @@ -199,7 +199,7 @@ func (i *IntelPMU) checkFileDescriptors() error { // soft limit for current process limit, err := i.fileInfo.fileLimit() if err != nil { - i.Log.Warnf("cannot obtain limit value of open files: %v", err) + i.Log.Warnf("Cannot obtain limit value of open files: %v", err) } else if allFd > limit { return fmt.Errorf("required file descriptors number `%d` exceeds soft limit of open files `%d`"+ ": consider increasing the limit", allFd, limit) @@ -249,7 +249,7 @@ func (i *IntelPMU) Stop() { } err := event.Deactivate() if err != nil { - i.Log.Warnf("failed to deactivate core event %q: %w", event, err) + i.Log.Warnf("Failed to deactivate core event %q: %v", event, err) } } } @@ -264,7 +264,7 @@ func (i *IntelPMU) Stop() { } err := event.Deactivate() if err != nil { - i.Log.Warnf("failed to deactivate uncore event %q: %w", event, err) + i.Log.Warnf("Failed to deactivate uncore event %q: %v", event, err) } } } diff --git a/plugins/inputs/intel_powerstat/intel_powerstat.go b/plugins/inputs/intel_powerstat/intel_powerstat.go index 1b4c8267a..cd9418976 100644 --- a/plugins/inputs/intel_powerstat/intel_powerstat.go +++ b/plugins/inputs/intel_powerstat/intel_powerstat.go @@ -133,7 +133,7 @@ func (p *PowerStat) Gather(acc telegraf.Accumulator) error { p.logOnce["msr"] = nil p.addPerCoreMetrics(acc) } else { - err := errors.New("error while trying to read MSR (probably msr module was not loaded)") + err := errors.New("Error while trying to read MSR (probably msr module was not loaded)") if val := p.logOnce["msr"]; val == nil || val.Error() != err.Error() { p.Log.Errorf("%v", err) // Remember that specific error occurs to omit logging next time @@ -171,7 +171,7 @@ func (p *PowerStat) addGlobalMetrics(acc telegraf.Accumulator) { if err != nil { // In case of an error skip calculating metrics for this socket if val := p.logOnce[socketID+"rapl"]; val == nil || val.Error() != err.Error() { - p.Log.Errorf("error fetching rapl data for socket %s, err: %v", socketID, err) + p.Log.Errorf("Error fetching rapl data for socket %s, err: %v", socketID, err) // Remember that specific error occurs for socketID to omit logging next time p.logOnce[socketID+"rapl"] = err } @@ -208,7 +208,7 @@ func maxDiePerSocket(_ string) int { func (p *PowerStat) addUncoreFreq(socketID string, die string, acc telegraf.Accumulator) { err := checkFile("/sys/devices/system/cpu/intel_uncore_frequency") if err != nil { - err := fmt.Errorf("error while checking existing intel_uncore_frequency (probably intel-uncore-frequency module was not loaded)") + err := fmt.Errorf("Error while checking existing intel_uncore_frequency (probably intel-uncore-frequency module was not loaded)") if val := p.logOnce["intel_uncore_frequency"]; val == nil || val.Error() != err.Error() { p.Log.Errorf("%v", err) // Remember that specific error occurs to omit logging next time @@ -228,18 +228,18 @@ func (p *PowerStat) readUncoreFreq(typeFreq string, socketID string, die string, p.logOnce[socketID+"msr"] = nil cpuID, err := p.GetCPUIDFromSocketID(socketID) if err != nil { - p.Log.Debugf("error while reading socket ID: %v", err) + p.Log.Debugf("Error while reading socket ID: %v", err) return } actualUncoreFreq, err := p.msr.readSingleMsr(cpuID, msrUncorePerfStatusString) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrUncorePerfStatusString, err) + p.Log.Debugf("Error while reading %s: %v", msrUncorePerfStatusString, err) return } actualUncoreFreq = (actualUncoreFreq & 0x3F) * 100 fields["uncore_frequency_mhz_cur"] = actualUncoreFreq } else { - err := errors.New("error while trying to read MSR (probably msr module was not loaded), uncore_frequency_mhz_cur metric will not be collected") + err := errors.New("Error while trying to read MSR (probably msr module was not loaded), uncore_frequency_mhz_cur metric will not be collected") if val := p.logOnce[socketID+"msr"]; val == nil || val.Error() != err.Error() { p.Log.Errorf("%v", err) // Remember that specific error occurs for socketID to omit logging next time @@ -249,12 +249,12 @@ func (p *PowerStat) readUncoreFreq(typeFreq string, socketID string, die string, } initMinFreq, err := p.msr.retrieveUncoreFrequency(socketID, typeFreq, "min", die) if err != nil { - p.Log.Errorf("error while retrieving minimum uncore frequency of the socket %s, err: %v", socketID, err) + p.Log.Errorf("Error while retrieving minimum uncore frequency of the socket %s, err: %v", socketID, err) return } initMaxFreq, err := p.msr.retrieveUncoreFrequency(socketID, typeFreq, "max", die) if err != nil { - p.Log.Errorf("error while retrieving maximum uncore frequency of the socket %s, err: %v", socketID, err) + p.Log.Errorf("Error while retrieving maximum uncore frequency of the socket %s, err: %v", socketID, err) return } @@ -272,7 +272,7 @@ func (p *PowerStat) readUncoreFreq(typeFreq string, socketID string, die string, func (p *PowerStat) addThermalDesignPowerMetric(socketID string, acc telegraf.Accumulator) { maxPower, err := p.rapl.getConstraintMaxPowerWatts(socketID) if err != nil { - p.Log.Errorf("error while retrieving TDP of the socket %s, err: %v", socketID, err) + p.Log.Errorf("Error while retrieving TDP of the socket %s, err: %v", socketID, err) return } @@ -334,7 +334,7 @@ func (p *PowerStat) addMetricsForSingleCore(cpuID string, acc telegraf.Accumulat err := p.msr.openAndReadMsr(cpuID) if err != nil { // In case of an error exit the function. All metrics past this point are dependent on MSR - p.Log.Debugf("error while reading msr: %v", err) + p.Log.Debugf("Error while reading msr: %v", err) return } } @@ -368,7 +368,7 @@ func (p *PowerStat) addCPUFrequencyMetric(cpuID string, acc telegraf.Accumulator // In case of an error leave func if err != nil { - p.Log.Debugf("error while reading file: %v", err) + p.Log.Debugf("Error while reading file: %v", err) return } @@ -435,7 +435,7 @@ func (p *PowerStat) addTurboRatioLimit(socketID string, acc telegraf.Accumulator } } if cpuID == "" || model == "" { - p.Log.Debug("error while reading socket ID") + p.Log.Debug("Error while reading socket ID") return } // dump_hsw_turbo_ratio_limit @@ -443,7 +443,7 @@ func (p *PowerStat) addTurboRatioLimit(socketID string, acc telegraf.Accumulator coreCounts := uint64(0x1211) // counting the number of active cores 17 and 18 msrTurboRatioLimit2, err := p.msr.readSingleMsr(cpuID, msrTurboRatioLimit2String) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrTurboRatioLimit2String, err) + p.Log.Debugf("Error while reading %s: %v", msrTurboRatioLimit2String, err) return } @@ -456,7 +456,7 @@ func (p *PowerStat) addTurboRatioLimit(socketID string, acc telegraf.Accumulator coreCounts := uint64(0x100F0E0D0C0B0A09) // counting the number of active cores 9 to 16 msrTurboRatioLimit1, err := p.msr.readSingleMsr(cpuID, msrTurboRatioLimit1String) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrTurboRatioLimit1String, err) + p.Log.Debugf("Error while reading %s: %v", msrTurboRatioLimit1String, err) return } calculateTurboRatioGroup(coreCounts, msrTurboRatioLimit1, turboRatioLimitGroups) @@ -478,14 +478,14 @@ func (p *PowerStat) addTurboRatioLimit(socketID string, acc telegraf.Accumulator coreCounts, err = p.msr.readSingleMsr(cpuID, msrTurboRatioLimit1String) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrTurboRatioLimit1String, err) + p.Log.Debugf("Error while reading %s: %v", msrTurboRatioLimit1String, err) return } } msrTurboRatioLimit, err := p.msr.readSingleMsr(cpuID, msrTurboRatioLimitString) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrTurboRatioLimitString, err) + p.Log.Debugf("Error while reading %s: %v", msrTurboRatioLimitString, err) return } calculateTurboRatioGroup(coreCounts, msrTurboRatioLimit, turboRatioLimitGroups) @@ -498,7 +498,7 @@ func (p *PowerStat) addTurboRatioLimit(socketID string, acc telegraf.Accumulator msrTurboRatioLimit, err := p.msr.readSingleMsr(cpuID, msrAtomCoreTurboRatiosString) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrAtomCoreTurboRatiosString, err) + p.Log.Debugf("Error while reading %s: %v", msrAtomCoreTurboRatiosString, err) return } value := uint64(0) @@ -515,7 +515,7 @@ func (p *PowerStat) addTurboRatioLimit(socketID string, acc telegraf.Accumulator if model == strconv.FormatInt(0x57, 10) { // INTEL_FAM6_XEON_PHI_KNL msrTurboRatioLimit, err := p.msr.readSingleMsr(cpuID, msrTurboRatioLimitString) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrTurboRatioLimitString, err) + p.Log.Debugf("Error while reading %s: %v", msrTurboRatioLimitString, err) return } @@ -556,7 +556,7 @@ func (p *PowerStat) addCPUBusyFrequencyMetric(cpuID string, acc telegraf.Accumul mperfDelta := coresData[cpuID].mperfDelta // Avoid division by 0 if mperfDelta == 0 { - p.Log.Errorf("mperf delta should not equal 0 on core %s", cpuID) + p.Log.Errorf("Value of mperf delta should not equal 0 on core %s", cpuID) return } aperfMperf := float64(coresData[cpuID].aperfDelta) / float64(mperfDelta) @@ -570,7 +570,7 @@ func (p *PowerStat) addCPUBusyFrequencyMetric(cpuID string, acc telegraf.Accumul } if interval == 0 { - p.Log.Errorf("interval between last two Telegraf cycles is 0") + p.Log.Errorf("Interval between last two Telegraf cycles is 0") return } @@ -594,7 +594,7 @@ func (p *PowerStat) addCPUC1StateResidencyMetric(cpuID string, acc telegraf.Accu timestampDeltaBig := new(big.Int).SetUint64(coresData[cpuID].timeStampCounterDelta) // Avoid division by 0 if timestampDeltaBig.Sign() < 1 { - p.Log.Errorf("timestamp delta value %v should not be lower than 1", timestampDeltaBig) + p.Log.Errorf("Timestamp delta value %v should not be lower than 1", timestampDeltaBig) return } @@ -633,7 +633,7 @@ func (p *PowerStat) addCPUC6StateResidencyMetric(cpuID string, acc telegraf.Accu coresData := p.msr.getCPUCoresData() // Avoid division by 0 if coresData[cpuID].timeStampCounterDelta == 0 { - p.Log.Errorf("timestamp counter on offset %s should not equal 0 on cpuID %s", + p.Log.Errorf("Timestamp counter on offset %d should not equal 0 on cpuID %s", timestampCounterLocation, cpuID) return } @@ -657,7 +657,7 @@ func (p *PowerStat) addCPUC0StateResidencyMetric(cpuID string, acc telegraf.Accu coresData := p.msr.getCPUCoresData() // Avoid division by 0 if coresData[cpuID].timeStampCounterDelta == 0 { - p.Log.Errorf("timestamp counter on offset %s should not equal 0 on cpuID %s", + p.Log.Errorf("Timestamp counter on offset %d should not equal 0 on cpuID %s", timestampCounterLocation, cpuID) return } @@ -686,13 +686,13 @@ func (p *PowerStat) addCPUC0StateResidencyMetric(cpuID string, acc telegraf.Accu func (p *PowerStat) addCPUBaseFreq(socketID string, acc telegraf.Accumulator) { cpuID, err := p.GetCPUIDFromSocketID(socketID) if err != nil { - p.Log.Debugf("error while getting CPU ID from Socket ID: %v", err) + p.Log.Debugf("Error while getting CPU ID from Socket ID: %v", err) return } msrPlatformInfoMsr, err := p.msr.readSingleMsr(cpuID, msrPlatformInfoString) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrPlatformInfoString, err) + p.Log.Debugf("Error while reading %s: %v", msrPlatformInfoString, err) return } @@ -700,7 +700,7 @@ func (p *PowerStat) addCPUBaseFreq(socketID string, acc telegraf.Accumulator) { // to get the freq -> ratio * busClock cpuBaseFreq := float64((msrPlatformInfoMsr>>8)&0xFF) * p.cpuBusClockValue if cpuBaseFreq == 0 { - p.Log.Debugf("error while adding CPU base frequency, cpuBaseFreq is zero for the socket: %s", socketID) + p.Log.Debugf("Error while adding CPU base frequency, cpuBaseFreq is zero for the socket: %s", socketID) return } @@ -716,7 +716,7 @@ func (p *PowerStat) addCPUBaseFreq(socketID string, acc telegraf.Accumulator) { func (p *PowerStat) getBusClock(cpuID string) float64 { cpuInfo, ok := p.cpuInfo[cpuID] if !ok { - p.Log.Debugf("cannot find cpuInfo for cpu: %s", cpuID) + p.Log.Debugf("Cannot find cpuInfo for cpu: %s", cpuID) return 0 } @@ -734,7 +734,7 @@ func (p *PowerStat) getBusClock(cpuID string) float64 { return p.getSilvermontBusClock(cpuID) } - p.Log.Debugf("couldn't find the freq for the model: %d", model) + p.Log.Debugf("Couldn't find the freq for the model: %s", model) return 0.0 } @@ -742,13 +742,13 @@ func (p *PowerStat) getSilvermontBusClock(cpuID string) float64 { silvermontFreqTable := []float64{83.3, 100.0, 133.3, 116.7, 80.0} msr, err := p.msr.readSingleMsr(cpuID, msrFSBFreqString) if err != nil { - p.Log.Debugf("error while reading %s: %v", msrFSBFreqString, err) + p.Log.Debugf("Error while reading %s: %v", msrFSBFreqString, err) return 0.0 } i := int(msr & 0xf) if i >= len(silvermontFreqTable) { - p.Log.Debugf("unknown msr value: %d, using default bus clock value: %d", i, silvermontFreqTable[3]) + p.Log.Debugf("Unknown msr value: %d, using default bus clock value: %f", i, silvermontFreqTable[3]) //same behaviour as in turbostat i = 3 } diff --git a/plugins/inputs/jti_openconfig_telemetry/jti_openconfig_telemetry.go b/plugins/inputs/jti_openconfig_telemetry/jti_openconfig_telemetry.go index af651ee70..8e8a202c9 100644 --- a/plugins/inputs/jti_openconfig_telemetry/jti_openconfig_telemetry.go +++ b/plugins/inputs/jti_openconfig_telemetry/jti_openconfig_telemetry.go @@ -327,7 +327,7 @@ func (m *OpenConfigTelemetry) collectData( if ok { timestamp = time.UnixMilli(int64(ts)) } else { - m.Log.Warnf("invalid type %T for _timestamp %v", group.data["_timestamp"], group.data["_timestamp"]) + m.Log.Warnf("Invalid type %T for _timestamp %v", group.data["_timestamp"], group.data["_timestamp"]) } } @@ -438,7 +438,7 @@ func (m *OpenConfigTelemetry) Start(acc telegraf.Accumulator) error { if m.Username != "" && m.Password != "" && m.ClientID != "" { if err := m.authenticate(ctx, server, grpcClientConn); err != nil { - m.Log.Errorf("error authenticating to %s: %w", grpcServer, err) + m.Log.Errorf("Error authenticating to %s: %v", grpcServer, err) continue } } diff --git a/plugins/inputs/kinesis_consumer/kinesis_consumer.go b/plugins/inputs/kinesis_consumer/kinesis_consumer.go index 3c9f763a9..1846aaf3b 100644 --- a/plugins/inputs/kinesis_consumer/kinesis_consumer.go +++ b/plugins/inputs/kinesis_consumer/kinesis_consumer.go @@ -145,14 +145,14 @@ func (k *KinesisConsumer) connect(ac telegraf.Accumulator) error { err := k.onMessage(k.acc, r) if err != nil { <-k.sem - k.Log.Errorf("Scan parser error: %s", err.Error()) + k.Log.Errorf("Scan parser error: %v", err) } return nil }) if err != nil { k.cancel() - k.Log.Errorf("Scan encountered an error: %s", err.Error()) + k.Log.Errorf("Scan encountered an error: %v", err) k.cons = nil } }() @@ -220,7 +220,7 @@ func (k *KinesisConsumer) onDelivery(ctx context.Context) { k.lastSeqNum = strToBint(sequenceNum) if err := k.checkpoint.SetCheckpoint(chk.streamName, chk.shardID, sequenceNum); err != nil { - k.Log.Debug("Setting checkpoint failed: %v", err) + k.Log.Debugf("Setting checkpoint failed: %v", err) } } else { k.Log.Debug("Metric group failed to process") diff --git a/plugins/inputs/mongodb/mongodb.go b/plugins/inputs/mongodb/mongodb.go index 745922f2c..21acdb0e4 100644 --- a/plugins/inputs/mongodb/mongodb.go +++ b/plugins/inputs/mongodb/mongodb.go @@ -139,7 +139,7 @@ func (m *MongoDB) setupConnection(connURL string) error { return fmt.Errorf("unable to ping MongoDB: %w", err) } - m.Log.Errorf("unable to ping MongoDB: %s", err) + m.Log.Errorf("Unable to ping MongoDB: %s", err) } server := &Server{ @@ -156,7 +156,7 @@ func (m *MongoDB) Stop() { for _, server := range m.clients { ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second) if err := server.client.Disconnect(ctx); err != nil { - m.Log.Errorf("disconnecting from %q failed: %s", server, err) + m.Log.Errorf("Disconnecting from %q failed: %v", server.hostname, err) } cancel() } @@ -172,14 +172,14 @@ func (m *MongoDB) Gather(acc telegraf.Accumulator) error { defer wg.Done() if m.DisconnectedServersBehavior == "skip" { if err := srv.ping(); err != nil { - m.Log.Debugf("failed to ping server: %s", err) + m.Log.Debugf("Failed to ping server: %s", err) return } } err := srv.gatherData(acc, m.GatherClusterStatus, m.GatherPerdbStats, m.GatherColStats, m.GatherTopStat, m.ColStatsDbs) if err != nil { - m.Log.Errorf("failed to gather data: %s", err) + m.Log.Errorf("Failed to gather data: %s", err) } }(client) } diff --git a/plugins/inputs/nfsclient/nfsclient.go b/plugins/inputs/nfsclient/nfsclient.go index 08cffa409..fc8ce474b 100644 --- a/plugins/inputs/nfsclient/nfsclient.go +++ b/plugins/inputs/nfsclient/nfsclient.go @@ -298,7 +298,7 @@ func (*NFSClient) SampleConfig() string { func (n *NFSClient) Gather(acc telegraf.Accumulator) error { file, err := os.Open(n.mountstatsPath) if err != nil { - n.Log.Errorf("Failed opening the [%s] file: %s ", file, err) + n.Log.Errorf("Failed opening the %q file: %v ", file.Name(), err) return err } defer file.Close() diff --git a/plugins/inputs/opensearch_query/opensearch_query.go b/plugins/inputs/opensearch_query/opensearch_query.go index 1a6a43552..8dcd87090 100644 --- a/plugins/inputs/opensearch_query/opensearch_query.go +++ b/plugins/inputs/opensearch_query/opensearch_query.go @@ -70,7 +70,7 @@ func (o *OpensearchQuery) Init() error { err := o.newClient() if err != nil { - o.Log.Errorf("error creating OpenSearch client: %w", err) + o.Log.Errorf("Error creating OpenSearch client: %v", err) } for i, agg := range o.Aggregations { diff --git a/plugins/inputs/proxmox/proxmox.go b/plugins/inputs/proxmox/proxmox.go index 07c4f0c20..1f4eff2e9 100644 --- a/plugins/inputs/proxmox/proxmox.go +++ b/plugins/inputs/proxmox/proxmox.go @@ -117,7 +117,7 @@ func gatherQemuData(px *Proxmox, acc telegraf.Accumulator) { func gatherVMData(px *Proxmox, acc telegraf.Accumulator, rt ResourceType) { vmStats, err := getVMStats(px, rt) if err != nil { - px.Log.Error("Error getting VM stats: %v", err) + px.Log.Errorf("Error getting VM stats: %v", err) return } diff --git a/plugins/inputs/socket_listener/stream_listener.go b/plugins/inputs/socket_listener/stream_listener.go index d50a8b594..960027c5e 100644 --- a/plugins/inputs/socket_listener/stream_listener.go +++ b/plugins/inputs/socket_listener/stream_listener.go @@ -117,15 +117,15 @@ func (l *streamListener) setupConnection(conn net.Conn) error { } if *l.KeepAlivePeriod == 0 { if err := tcpConn.SetKeepAlive(false); err != nil { - l.Log.Warnf("Cannot set keep-alive: %w", err) + l.Log.Warnf("Cannot set keep-alive: %v", err) } } else { if err := tcpConn.SetKeepAlive(true); err != nil { - l.Log.Warnf("Cannot set keep-alive: %w", err) + l.Log.Warnf("Cannot set keep-alive: %v", err) } err := tcpConn.SetKeepAlivePeriod(time.Duration(*l.KeepAlivePeriod)) if err != nil { - l.Log.Warnf("Cannot set keep-alive period: %w", err) + l.Log.Warnf("Cannot set keep-alive period: %v", err) } } } diff --git a/plugins/inputs/upsd/upsd.go b/plugins/inputs/upsd/upsd.go index 8e3c55ee6..1f467c0f2 100644 --- a/plugins/inputs/upsd/upsd.go +++ b/plugins/inputs/upsd/upsd.go @@ -67,12 +67,12 @@ func (u *Upsd) gatherUps(acc telegraf.Accumulator, name string, variables []nut. timeLeftS, err := internal.ToFloat64(metrics["battery.runtime"]) if err != nil { - u.Log.Warnf("'battery.runtime' type is not supported: %w", err) + u.Log.Warnf("Type for 'battery.runtime' is not supported: %v", err) } timeLeftNS, err := internal.ToInt64(timeLeftS * 1_000_000_000) if err != nil { - u.Log.Warnf("converting 'battery.runtime' to 'time_left_ns' failed: %w", err) + u.Log.Warnf("Converting 'battery.runtime' to 'time_left_ns' failed: %v", err) } fields := map[string]interface{}{ diff --git a/plugins/inputs/win_eventlog/win_eventlog.go b/plugins/inputs/win_eventlog/win_eventlog.go index 142bcf9ba..7515966ef 100644 --- a/plugins/inputs/win_eventlog/win_eventlog.go +++ b/plugins/inputs/win_eventlog/win_eventlog.go @@ -89,7 +89,7 @@ func (w *WinEventLog) Stop() { func (w *WinEventLog) GetState() interface{} { bookmarkXML, err := w.renderBookmark(w.bookmark) if err != nil { - w.Log.Errorf("State-persistence failed, cannot render bookmark: %w", err) + w.Log.Errorf("State-persistence failed, cannot render bookmark: %v", err) return "" } return bookmarkXML diff --git a/plugins/inputs/win_perf_counters/win_perf_counters.go b/plugins/inputs/win_perf_counters/win_perf_counters.go index ed70c262d..37c3bec07 100644 --- a/plugins/inputs/win_perf_counters/win_perf_counters.go +++ b/plugins/inputs/win_perf_counters/win_perf_counters.go @@ -380,7 +380,7 @@ func (m *WinPerfCounters) ParseConfig() error { PerfObject.Measurement, PerfObject.IncludeTotal, PerfObject.UseRawValues) if err != nil { if PerfObject.FailOnMissing || PerfObject.WarnOnMissing { - m.Log.Errorf("invalid counterPath %q: %s", counterPath, err.Error()) + m.Log.Errorf("Invalid counterPath %q: %s", counterPath, err.Error()) } if PerfObject.FailOnMissing { return err @@ -449,10 +449,10 @@ func (m *WinPerfCounters) Gather(acc telegraf.Accumulator) error { for _, hostCounterInfo := range m.hostCounters { wg.Add(1) go func(hostInfo *hostCountersInfo) { - m.Log.Debugf("gathering from %s", hostInfo.computer) + m.Log.Debugf("Gathering from %s", hostInfo.computer) start := time.Now() err := m.gatherComputerCounters(hostInfo, acc) - m.Log.Debugf("gathering from %s finished in %.3fs", hostInfo.computer, time.Since(start)) + m.Log.Debugf("Gathering from %s finished in %v", hostInfo.computer, time.Since(start)) if err != nil { acc.AddError(fmt.Errorf("error during collecting data on host %q: %w", hostInfo.computer, err)) } @@ -482,7 +482,7 @@ func (m *WinPerfCounters) gatherComputerCounters(hostCounterInfo *hostCountersIn if !isKnownCounterDataError(err) { return fmt.Errorf("error while getting value for counter %q: %w", metric.counterPath, err) } - m.Log.Warnf("error while getting value for counter %q, instance: %s, will skip metric: %v", metric.counterPath, metric.instance, err) + m.Log.Warnf("Error while getting value for counter %q, instance: %s, will skip metric: %v", metric.counterPath, metric.instance, err) continue } addCounterMeasurement(metric, metric.instance, value, collectedFields) @@ -498,7 +498,7 @@ func (m *WinPerfCounters) gatherComputerCounters(hostCounterInfo *hostCountersIn if !isKnownCounterDataError(err) { return fmt.Errorf("error while getting value for counter %q: %w", metric.counterPath, err) } - m.Log.Warnf("error while getting value for counter %q, instance: %s, will skip metric: %v", metric.counterPath, metric.instance, err) + m.Log.Warnf("Error while getting value for counter %q, instance: %s, will skip metric: %v", metric.counterPath, metric.instance, err) continue } for _, cValue := range counterValues { @@ -589,14 +589,14 @@ func (m *WinPerfCounters) Init() error { for _, wildcard := range wildcards { if strings.Contains(object.ObjectName, wildcard) { found = true - m.Log.Errorf("object: %s, contains wildcard %s", object.ObjectName, wildcard) + m.Log.Errorf("Object: %s, contains wildcard %s", object.ObjectName, wildcard) } } for _, counter := range object.Counters { for _, wildcard := range wildcards { if strings.Contains(counter, wildcard) { found = true - m.Log.Errorf("object: %s, counter: %s contains wildcard %s", object.ObjectName, counter, wildcard) + m.Log.Errorf("Object: %s, counter: %s contains wildcard %s", object.ObjectName, counter, wildcard) } } } diff --git a/plugins/outputs/cloud_pubsub/cloud_pubsub.go b/plugins/outputs/cloud_pubsub/cloud_pubsub.go index 18f58c3c7..764448b63 100644 --- a/plugins/outputs/cloud_pubsub/cloud_pubsub.go +++ b/plugins/outputs/cloud_pubsub/cloud_pubsub.go @@ -189,7 +189,7 @@ func (ps *PubSub) toMessages(metrics []telegraf.Metric) ([]*pubsub.Message, erro b, err = ps.compressData(b) if err != nil { - ps.Log.Errorf("unable to compress message with %s: %w", ps.ContentEncoding, err) + ps.Log.Errorf("Unable to compress message with %s: %v", ps.ContentEncoding, err) continue } diff --git a/plugins/outputs/mqtt/mqtt.go b/plugins/outputs/mqtt/mqtt.go index a82db629f..258673493 100644 --- a/plugins/outputs/mqtt/mqtt.go +++ b/plugins/outputs/mqtt/mqtt.go @@ -167,7 +167,7 @@ func (m *MQTT) Write(metrics []telegraf.Metric) error { for _, msg := range topicMessages { if err := m.client.Publish(msg.topic, msg.payload); err != nil { - m.Log.Warn("Could not publish message to MQTT server, %s", err) + m.Log.Warnf("Could not publish message to MQTT server: %v", err) } } @@ -179,7 +179,7 @@ func (m *MQTT) collectNonBatch(hostname string, metrics []telegraf.Metric) []mes for _, metric := range metrics { topic, err := m.generator.Generate(hostname, metric) if err != nil { - m.Log.Warnf("Generating topic name failed: %w", err) + m.Log.Warnf("Generating topic name failed: %v", err) m.Log.Debugf("metric was: %v", metric) continue } @@ -201,7 +201,7 @@ func (m *MQTT) collectBatch(hostname string, metrics []telegraf.Metric) []messag for _, metric := range metrics { topic, err := m.generator.Generate(hostname, metric) if err != nil { - m.Log.Warnf("Generating topic name failed: %w", err) + m.Log.Warnf("Generating topic name failed: %v", err) m.Log.Debugf("metric was: %v", metric) continue } @@ -225,7 +225,7 @@ func (m *MQTT) collectField(hostname string, metrics []telegraf.Metric) []messag for _, metric := range metrics { topic, err := m.generator.Generate(hostname, metric) if err != nil { - m.Log.Warnf("Generating topic name failed: %w", err) + m.Log.Warnf("Generating topic name failed: %v", err) m.Log.Debugf("metric was: %v", metric) continue } @@ -249,14 +249,14 @@ func (m *MQTT) collectHomieV4(hostname string, metrics []telegraf.Metric) []mess for _, metric := range metrics { topic, err := m.generator.Generate(hostname, metric) if err != nil { - m.Log.Warnf("Generating topic name failed: %w", err) + m.Log.Warnf("Generating topic name failed: %v", err) m.Log.Debugf("metric was: %v", metric) continue } msgs, nodeID, err := m.collectHomieDeviceMessages(topic, metric) if err != nil { - m.Log.Warnf(err.Error()) + m.Log.Warn(err.Error()) m.Log.Debugf("metric was: %v", metric) continue } diff --git a/plugins/outputs/nebius_cloud_monitoring/nebius_cloud_monitoring.go b/plugins/outputs/nebius_cloud_monitoring/nebius_cloud_monitoring.go index 0e4f9ca6b..2673f0533 100644 --- a/plugins/outputs/nebius_cloud_monitoring/nebius_cloud_monitoring.go +++ b/plugins/outputs/nebius_cloud_monitoring/nebius_cloud_monitoring.go @@ -134,7 +134,7 @@ func (a *NebiusCloudMonitoring) Write(metrics []telegraf.Metric) error { for _, field := range m.FieldList() { value, err := internal.ToFloat64(field.Value) if err != nil { - a.Log.Errorf("skipping value: %w", err.Error()) + a.Log.Errorf("Skipping value: %v", err) continue } diff --git a/plugins/outputs/opentelemetry/opentelemetry.go b/plugins/outputs/opentelemetry/opentelemetry.go index 7b5a92d26..ae0ec9e49 100644 --- a/plugins/outputs/opentelemetry/opentelemetry.go +++ b/plugins/outputs/opentelemetry/opentelemetry.go @@ -139,7 +139,7 @@ func (o *OpenTelemetry) Write(metrics []telegraf.Metric) error { // sort the timestamps we collected sort.Slice(timestamps, func(i, j int) bool { return timestamps[i] < timestamps[j] }) - o.Log.Debugf("received %d metrics and split into %d groups by timestamp", len(metrics), len(metricBatch)) + o.Log.Debugf("Received %d metrics and split into %d groups by timestamp", len(metrics), len(metricBatch)) for _, timestamp := range timestamps { if err := o.sendBatch(metricBatch[timestamp]); err != nil { return err @@ -165,12 +165,12 @@ func (o *OpenTelemetry) sendBatch(metrics []telegraf.Metric) error { case telegraf.Summary: vType = common.InfluxMetricValueTypeSummary default: - o.Log.Warnf("unrecognized metric type %Q", metric.Type()) + o.Log.Warnf("Unrecognized metric type %v", metric.Type()) continue } err := batch.AddPoint(metric.Name(), metric.Tags(), metric.Fields(), metric.Time(), vType) if err != nil { - o.Log.Warnf("failed to add point: %s", err) + o.Log.Warnf("Failed to add point: %v", err) continue } } diff --git a/plugins/outputs/postgresql/table_manager.go b/plugins/outputs/postgresql/table_manager.go index 3e3171ecc..e240cc3a3 100644 --- a/plugins/outputs/postgresql/table_manager.go +++ b/plugins/outputs/postgresql/table_manager.go @@ -89,7 +89,7 @@ func (tm *TableManager) MatchSource(ctx context.Context, db dbh, rowSource *Tabl if isTempError(err) { return err } - tm.Postgresql.Logger.Errorf("permanent error updating schema for %s: %w", tagTable.name, err) + tm.Postgresql.Logger.Errorf("Permanent error updating schema for %s: %v", tagTable.name, err) } if len(missingCols) > 0 { @@ -100,7 +100,7 @@ func (tm *TableManager) MatchSource(ctx context.Context, db dbh, rowSource *Tabl } colDefs = append(colDefs, col.Name+" "+col.Type) } - tm.Logger.Errorf("table %q is missing tag columns (dropping metrics): %s", + tm.Logger.Errorf("Table %q is missing tag columns (dropping metrics): %s", tagTable.name, strings.Join(colDefs, ", ")) } @@ -120,7 +120,7 @@ func (tm *TableManager) MatchSource(ctx context.Context, db dbh, rowSource *Tabl if isTempError(err) { return err } - tm.Postgresql.Logger.Errorf("permanent error updating schema for %s: %w", metricTable.name, err) + tm.Postgresql.Logger.Errorf("Permanent error updating schema for %s: %v", metricTable.name, err) } if len(missingCols) > 0 { @@ -131,7 +131,7 @@ func (tm *TableManager) MatchSource(ctx context.Context, db dbh, rowSource *Tabl } colDefs = append(colDefs, col.Name+" "+col.Type) } - tm.Logger.Errorf("table %q is missing columns (omitting fields): %s", + tm.Logger.Errorf("Table %q is missing columns (omitting fields): %s", metricTable.name, strings.Join(colDefs, ", ")) } @@ -189,7 +189,7 @@ func (tm *TableManager) EnsureStructure( if col.Role == utils.TagColType { return nil, fmt.Errorf("column name too long: %q", col.Name) } - tm.Postgresql.Logger.Errorf("column name too long: %q", col.Name) + tm.Postgresql.Logger.Errorf("Column name too long: %q", col.Name) invalidColumns = append(invalidColumns, col) } diff --git a/plugins/outputs/timestream/timestream.go b/plugins/outputs/timestream/timestream.go index f5fe7ab99..0d6762fab 100644 --- a/plugins/outputs/timestream/timestream.go +++ b/plugins/outputs/timestream/timestream.go @@ -200,7 +200,7 @@ func (t *Timestream) Connect() error { t.Log.Errorf("Couldn't describe database %q. Check error, fix permissions, connectivity, create database.", t.DatabaseName) return err } - t.Log.Infof("Describe database %q returned %q.", t.DatabaseName, describeDatabaseOutput) + t.Log.Infof("Describe database %q returned %v", t.DatabaseName, describeDatabaseOutput) } t.svc = svc diff --git a/plugins/outputs/wavefront/wavefront.go b/plugins/outputs/wavefront/wavefront.go index 05c38253d..5ef8991b4 100644 --- a/plugins/outputs/wavefront/wavefront.go +++ b/plugins/outputs/wavefront/wavefront.go @@ -88,15 +88,15 @@ func (w *Wavefront) Connect() error { } var connectionURL string if w.URL != "" { - w.Log.Debug("connecting over http/https using Url: %s", w.URL) + w.Log.Debugf("Connecting over http/https using url: %s", w.URL) connectionURLWithToken, err := w.senderURLFromURLAndToken() if err != nil { return err } connectionURL = connectionURLWithToken } else { - w.Log.Warnf("configuration with host/port is deprecated. Please use url.") - w.Log.Debugf("connecting over http using Host: %q and Port: %d", w.Host, w.Port) + w.Log.Warn("Configuration with host/port is deprecated. Please use url.") + w.Log.Debugf("Connecting over http using Host: %q and Port: %d", w.Host, w.Port) connectionURL = senderURLFromHostAndPort(w.Host, w.Port) } @@ -147,8 +147,8 @@ func (w *Wavefront) Write(metrics []telegraf.Metric) error { } } } - w.Log.Errorf("non-retryable error during Wavefront.Write: %v", err) - w.Log.Debugf("non-retryable metric data: %+v", point) + w.Log.Errorf("Non-retryable error during Wavefront.Write: %v", err) + w.Log.Debugf("Non-retryable metric data: %+v", point) } } } diff --git a/plugins/outputs/yandex_cloud_monitoring/yandex_cloud_monitoring.go b/plugins/outputs/yandex_cloud_monitoring/yandex_cloud_monitoring.go index 10c3d43a6..f0b8a8114 100644 --- a/plugins/outputs/yandex_cloud_monitoring/yandex_cloud_monitoring.go +++ b/plugins/outputs/yandex_cloud_monitoring/yandex_cloud_monitoring.go @@ -126,7 +126,7 @@ func (a *YandexCloudMonitoring) Write(metrics []telegraf.Metric) error { for _, field := range m.FieldList() { value, err := internal.ToFloat64(field.Value) if err != nil { - a.Log.Errorf("skipping value: %w", err.Error()) + a.Log.Errorf("Skipping value: %v", err) continue } @@ -180,7 +180,7 @@ func getResponseFromMetadata(c *http.Client, metadataURL string) ([]byte, error) } func (a *YandexCloudMonitoring) getFolderIDFromMetadata() (string, error) { - a.Log.Infof("getting folder ID in %s", a.MetadataFolderURL) + a.Log.Infof("Getting folder ID in %s", a.MetadataFolderURL) body, err := getResponseFromMetadata(a.client, a.MetadataFolderURL) if err != nil { return "", err @@ -193,7 +193,7 @@ func (a *YandexCloudMonitoring) getFolderIDFromMetadata() (string, error) { } func (a *YandexCloudMonitoring) getIAMTokenFromMetadata() (string, int, error) { - a.Log.Debugf("getting new IAM token in %s", a.MetadataTokenURL) + a.Log.Debugf("Getting new IAM token in %s", a.MetadataTokenURL) body, err := getResponseFromMetadata(a.client, a.MetadataTokenURL) if err != nil { return "", 0, err @@ -230,7 +230,7 @@ func (a *YandexCloudMonitoring) send(body []byte) error { } req.Header.Set("Authorization", "Bearer "+a.IAMToken) - a.Log.Debugf("sending metrics to %s", req.URL.String()) + a.Log.Debugf("Sending metrics to %s", req.URL.String()) a.Log.Debugf("body: %s", body) resp, err := a.client.Do(req) if err != nil { diff --git a/plugins/processors/scale/scale.go b/plugins/processors/scale/scale.go index 649864040..38333f789 100644 --- a/plugins/processors/scale/scale.go +++ b/plugins/processors/scale/scale.go @@ -46,7 +46,8 @@ func (s *Scaling) Init() error { anyMinMaxSet := s.OutMax != nil || s.OutMin != nil || s.InMax != nil || s.InMin != nil factorSet := s.Factor != nil || s.Offset != nil if anyMinMaxSet && factorSet { - return fmt.Errorf("cannot use factor/offset and minimum/maximum at the same time for fields %s", strings.Join(s.Fields, ",")) + return fmt.Errorf("cannot use factor/offset and minimum/maximum at the same time for fields %s", + strings.Join(s.Fields, ",")) } else if anyMinMaxSet && !allMinMaxSet { return fmt.Errorf("all minimum and maximum values need to be set for fields %s", strings.Join(s.Fields, ",")) } else if !anyMinMaxSet && !factorSet { @@ -96,7 +97,7 @@ func (s *Scale) Init() error { for _, field := range s.Scalings[i].Fields { // only generate a warning for the first duplicate field filter if warn, ok := allFields[field]; ok && warn { - s.Log.Warnf("filter field %q used twice in scalings", field) + s.Log.Warnf("Filter field %q used twice in scalings", field) allFields[field] = false } else { allFields[field] = true @@ -122,7 +123,7 @@ func (s *Scale) scaleValues(metric telegraf.Metric) { v, err := internal.ToFloat64(field.Value) if err != nil { - s.Log.Errorf("error converting %q to float: %w\n", field.Key, err) + s.Log.Errorf("Error converting %q to float: %v", field.Key, err) continue } diff --git a/plugins/serializers/cloudevents/cloudevents.go b/plugins/serializers/cloudevents/cloudevents.go index fed1d42b3..0112d9fe4 100644 --- a/plugins/serializers/cloudevents/cloudevents.go +++ b/plugins/serializers/cloudevents/cloudevents.go @@ -139,7 +139,7 @@ func (s *Serializer) batchEvents(metrics []telegraf.Metric) ([]byte, error) { for _, m := range metrics { e, err := s.createEvent(m) if err != nil { - s.Log.Errorf("creating event for %v failed: %w", m, err) + s.Log.Errorf("Creating event for %v failed: %v", m, err) continue } events = append(events, e)