chore(linters): replace 'fmt.Sprintf' with string addition (#14699)

This commit is contained in:
Paweł Żak 2024-02-08 21:46:22 +01:00 committed by GitHub
parent 19fd5712ae
commit 98ec91478a
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
63 changed files with 198 additions and 205 deletions

View File

@ -1,7 +1,6 @@
package main package main
import ( import (
"fmt"
"log" "log"
"net/http" "net/http"
"strings" "strings"
@ -28,7 +27,7 @@ func (p *PprofServer) Start(address string) {
pprofHostPort := address pprofHostPort := address
parts := strings.Split(pprofHostPort, ":") parts := strings.Split(pprofHostPort, ":")
if len(parts) == 2 && parts[0] == "" { if len(parts) == 2 && parts[0] == "" {
pprofHostPort = fmt.Sprintf("localhost:%s", parts[1]) pprofHostPort = "localhost:" + parts[1]
} }
pprofHostPort = "http://" + pprofHostPort + "/debug/pprof" pprofHostPort = "http://" + pprofHostPort + "/debug/pprof"

View File

@ -99,7 +99,7 @@ func testBuildTags(t *testing.T, buildComment string, pluginCategory string, plu
// For ex ["!custom", "inputs", "inputs.docker"] returns "inputs.docker" // For ex ["!custom", "inputs", "inputs.docker"] returns "inputs.docker"
func getPluginBuildTag(tags []string, pluginCategory string) string { func getPluginBuildTag(tags []string, pluginCategory string) string {
for _, tag := range tags { for _, tag := range tags {
if strings.HasPrefix(tag, fmt.Sprintf("%s.", pluginCategory)) { if strings.HasPrefix(tag, pluginCategory+".") {
return tag return tag
} }
} }

View File

@ -41,10 +41,10 @@ func generateCert(host string, rsaBits int, certFile, keyFile string, dur time.D
rsaBits = 2048 rsaBits = 2048
} }
if len(certFile) == 0 { if len(certFile) == 0 {
certFile = fmt.Sprintf("%s/cert.pem", dir) certFile = dir + "/cert.pem"
} }
if len(keyFile) == 0 { if len(keyFile) == 0 {
keyFile = fmt.Sprintf("%s/key.pem", dir) keyFile = dir + "/key.pem"
} }
priv, err := rsa.GenerateKey(rand.Reader, rsaBits) priv, err := rsa.GenerateKey(rand.Reader, rsaBits)

View File

@ -307,7 +307,7 @@ func (a *Aerospike) getSetInfo(namespaceSet string, n *as.Node, infoPolicy *as.I
func (a *Aerospike) parseSetInfo(acc telegraf.Accumulator, stats map[string]string, hostPort string, namespaceSet string, nodeName string) { func (a *Aerospike) parseSetInfo(acc telegraf.Accumulator, stats map[string]string, hostPort string, namespaceSet string, nodeName string) {
stat := strings.Split( stat := strings.Split(
strings.TrimSuffix( strings.TrimSuffix(
stats[fmt.Sprintf("sets/%s", namespaceSet)], ";"), ":") stats["sets/"+namespaceSet], ";"), ":")
nTags := map[string]string{ nTags := map[string]string{
"aerospike_host": hostPort, "aerospike_host": hostPort,
"node_name": nodeName, "node_name": nodeName,

View File

@ -484,7 +484,7 @@ func formatField(metricName string, statistic string) string {
func formatMeasurement(project string) string { func formatMeasurement(project string) string {
project = strings.ReplaceAll(project, "/", "_") project = strings.ReplaceAll(project, "/", "_")
project = snakeCase(project) project = snakeCase(project)
return fmt.Sprintf("aliyuncms_%s", project) return "aliyuncms_" + project
} }
func snakeCase(s string) string { func snakeCase(s string) string {

View File

@ -1,7 +1,6 @@
package aurora package aurora
import ( import (
"fmt"
"net/http" "net/http"
"net/http/httptest" "net/http/httptest"
"net/url" "net/url"
@ -21,7 +20,7 @@ func TestAurora(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -220,7 +219,7 @@ func TestBasicAuth(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {

View File

@ -10,10 +10,11 @@ import (
"path/filepath" "path/filepath"
"strings" "strings"
"golang.org/x/sys/unix"
"github.com/influxdata/telegraf" "github.com/influxdata/telegraf"
"github.com/influxdata/telegraf/filter" "github.com/influxdata/telegraf/filter"
"github.com/influxdata/telegraf/plugins/inputs/system" "github.com/influxdata/telegraf/plugins/inputs/system"
"golang.org/x/sys/unix"
) )
type DiskIO struct { type DiskIO struct {
@ -68,7 +69,7 @@ func (d *DiskIO) diskInfo(devName string) (map[string]string, error) {
_, err := os.Stat(udevDataPath) _, err := os.Stat(udevDataPath)
if err != nil { if err != nil {
// This path failed, try the fallback .udev style (non-systemd) // This path failed, try the fallback .udev style (non-systemd)
udevDataPath = fmt.Sprintf("/dev/.udev/db/block:%s", devName) udevDataPath = "/dev/.udev/db/block:" + devName
_, err := os.Stat(udevDataPath) _, err := os.Stat(udevDataPath)
if err != nil { if err != nil {
// Giving up, cannot retrieve disk info // Giving up, cannot retrieve disk info

View File

@ -6,8 +6,9 @@ import (
"net" "net"
"testing" "testing"
"github.com/influxdata/telegraf/testutil"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
"github.com/influxdata/telegraf/testutil"
) )
func TestDisqueGeneratesMetricsIntegration(t *testing.T) { func TestDisqueGeneratesMetricsIntegration(t *testing.T) {
@ -47,7 +48,7 @@ func TestDisqueGeneratesMetricsIntegration(t *testing.T) {
} }
}() }()
addr := fmt.Sprintf("disque://%s", l.Addr().String()) addr := "disque://" + l.Addr().String()
r := &Disque{ r := &Disque{
Servers: []string{addr}, Servers: []string{addr},
@ -117,7 +118,7 @@ func TestDisqueCanPullStatsFromMultipleServersIntegration(t *testing.T) {
} }
}() }()
addr := fmt.Sprintf("disque://%s", l.Addr().String()) addr := "disque://" + l.Addr().String()
r := &Disque{ r := &Disque{
Servers: []string{addr}, Servers: []string{addr},

View File

@ -792,7 +792,7 @@ func getDeviceStatMap(blkioStats types.BlkioStats) map[string]map[string]interfa
deviceStatMap[device] = make(map[string]interface{}) deviceStatMap[device] = make(map[string]interface{})
} }
field := fmt.Sprintf("io_service_bytes_recursive_%s", strings.ToLower(metric.Op)) field := "io_service_bytes_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
@ -803,31 +803,31 @@ func getDeviceStatMap(blkioStats types.BlkioStats) map[string]map[string]interfa
deviceStatMap[device] = make(map[string]interface{}) deviceStatMap[device] = make(map[string]interface{})
} }
field := fmt.Sprintf("io_serviced_recursive_%s", strings.ToLower(metric.Op)) field := "io_serviced_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoQueuedRecursive { for _, metric := range blkioStats.IoQueuedRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_queue_recursive_%s", strings.ToLower(metric.Op)) field := "io_queue_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoServiceTimeRecursive { for _, metric := range blkioStats.IoServiceTimeRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_service_time_recursive_%s", strings.ToLower(metric.Op)) field := "io_service_time_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoWaitTimeRecursive { for _, metric := range blkioStats.IoWaitTimeRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_wait_time_%s", strings.ToLower(metric.Op)) field := "io_wait_time_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoMergedRecursive { for _, metric := range blkioStats.IoMergedRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_merged_recursive_%s", strings.ToLower(metric.Op)) field := "io_merged_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }

View File

@ -90,7 +90,7 @@ func (d *Dovecot) gatherServer(addr string, acc telegraf.Accumulator, qtype stri
return fmt.Errorf("setting deadline failed for dovecot server %q: %w", addr, err) return fmt.Errorf("setting deadline failed for dovecot server %q: %w", addr, err)
} }
msg := fmt.Sprintf("EXPORT\t%s", qtype) msg := "EXPORT\t" + qtype
if len(filter) > 0 { if len(filter) > 0 {
msg += fmt.Sprintf("\t%s=%s", qtype, filter) msg += fmt.Sprintf("\t%s=%s", qtype, filter)
} }

View File

@ -18,7 +18,7 @@ func Test_LinkStatusCommand(t *testing.T) {
response := fmt.Sprintf(`{%q:{%q: "DOWN"}}`, ethdevLinkStatusCommand, linkStatusStringFieldName) response := fmt.Sprintf(`{%q:{%q: "DOWN"}}`, ethdevLinkStatusCommand, linkStatusStringFieldName)
simulateResponse(mockConn, response, nil) simulateResponse(mockConn, response, nil)
dpdkConn := dpdk.connectors[0] dpdkConn := dpdk.connectors[0]
dpdkConn.processCommand(mockAcc, testutil.Logger{}, fmt.Sprintf("%s,1", ethdevLinkStatusCommand), nil) dpdkConn.processCommand(mockAcc, testutil.Logger{}, ethdevLinkStatusCommand+",1", nil)
expected := []telegraf.Metric{ expected := []telegraf.Metric{
testutil.MustMetric( testutil.MustMetric(
@ -45,7 +45,7 @@ func Test_LinkStatusCommand(t *testing.T) {
response := fmt.Sprintf(`{%q:{%q: "UP"}}`, ethdevLinkStatusCommand, linkStatusStringFieldName) response := fmt.Sprintf(`{%q:{%q: "UP"}}`, ethdevLinkStatusCommand, linkStatusStringFieldName)
simulateResponse(mockConn, response, nil) simulateResponse(mockConn, response, nil)
dpdkConn := dpdk.connectors[0] dpdkConn := dpdk.connectors[0]
dpdkConn.processCommand(mockAcc, testutil.Logger{}, fmt.Sprintf("%s,1", ethdevLinkStatusCommand), nil) dpdkConn.processCommand(mockAcc, testutil.Logger{}, ethdevLinkStatusCommand+",1", nil)
expected := []telegraf.Metric{ expected := []telegraf.Metric{
testutil.MustMetric( testutil.MustMetric(
@ -72,7 +72,7 @@ func Test_LinkStatusCommand(t *testing.T) {
response := fmt.Sprintf(`{%q:{}}`, ethdevLinkStatusCommand) response := fmt.Sprintf(`{%q:{}}`, ethdevLinkStatusCommand)
simulateResponse(mockConn, response, nil) simulateResponse(mockConn, response, nil)
dpdkConn := dpdk.connectors[0] dpdkConn := dpdk.connectors[0]
dpdkConn.processCommand(mockAcc, testutil.Logger{}, fmt.Sprintf("%s,1", ethdevLinkStatusCommand), nil) dpdkConn.processCommand(mockAcc, testutil.Logger{}, ethdevLinkStatusCommand+",1", nil)
actual := mockAcc.GetTelegrafMetrics() actual := mockAcc.GetTelegrafMetrics()
testutil.RequireMetricsEqual(t, nil, actual, testutil.IgnoreTime()) testutil.RequireMetricsEqual(t, nil, actual, testutil.IgnoreTime())
@ -84,7 +84,7 @@ func Test_LinkStatusCommand(t *testing.T) {
response := fmt.Sprintf(`{%q:{"tag1": 1}}`, ethdevLinkStatusCommand) response := fmt.Sprintf(`{%q:{"tag1": 1}}`, ethdevLinkStatusCommand)
simulateResponse(mockConn, response, nil) simulateResponse(mockConn, response, nil)
dpdkConn := dpdk.connectors[0] dpdkConn := dpdk.connectors[0]
dpdkConn.processCommand(mockAcc, testutil.Logger{}, fmt.Sprintf("%s,1", ethdevLinkStatusCommand), nil) dpdkConn.processCommand(mockAcc, testutil.Logger{}, ethdevLinkStatusCommand+",1", nil)
expected := []telegraf.Metric{ expected := []telegraf.Metric{
testutil.MustMetric( testutil.MustMetric(
"dpdk", "dpdk",
@ -109,7 +109,7 @@ func Test_LinkStatusCommand(t *testing.T) {
response := fmt.Sprintf(`{%q:{%q: "BOB"}}`, ethdevLinkStatusCommand, linkStatusStringFieldName) response := fmt.Sprintf(`{%q:{%q: "BOB"}}`, ethdevLinkStatusCommand, linkStatusStringFieldName)
simulateResponse(mockConn, response, nil) simulateResponse(mockConn, response, nil)
dpdkConn := dpdk.connectors[0] dpdkConn := dpdk.connectors[0]
dpdkConn.processCommand(mockAcc, testutil.Logger{}, fmt.Sprintf("%s,1", ethdevLinkStatusCommand), nil) dpdkConn.processCommand(mockAcc, testutil.Logger{}, ethdevLinkStatusCommand+",1", nil)
expected := []telegraf.Metric{ expected := []telegraf.Metric{
testutil.MustMetric( testutil.MustMetric(

View File

@ -46,7 +46,7 @@ func isInMemorySocketPath(filePath, socketPath string) bool {
return true return true
} }
socketPathPrefix := fmt.Sprintf("%s:", socketPath) socketPathPrefix := socketPath + ":"
if strings.HasPrefix(filePath, socketPathPrefix) { if strings.HasPrefix(filePath, socketPathPrefix) {
suffix := filePath[len(socketPathPrefix):] suffix := filePath[len(socketPathPrefix):]
if number, err := strconv.Atoi(suffix); err == nil { if number, err := strconv.Atoi(suffix); err == nil {

View File

@ -207,7 +207,7 @@ func blkstats(id string, stats *types.StatsJSON, acc telegraf.Accumulator, tags
deviceStatMap[device] = make(map[string]interface{}) deviceStatMap[device] = make(map[string]interface{})
} }
field := fmt.Sprintf("io_service_bytes_recursive_%s", strings.ToLower(metric.Op)) field := "io_service_bytes_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
@ -218,31 +218,31 @@ func blkstats(id string, stats *types.StatsJSON, acc telegraf.Accumulator, tags
deviceStatMap[device] = make(map[string]interface{}) deviceStatMap[device] = make(map[string]interface{})
} }
field := fmt.Sprintf("io_serviced_recursive_%s", strings.ToLower(metric.Op)) field := "io_serviced_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoQueuedRecursive { for _, metric := range blkioStats.IoQueuedRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_queue_recursive_%s", strings.ToLower(metric.Op)) field := "io_queue_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoServiceTimeRecursive { for _, metric := range blkioStats.IoServiceTimeRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_service_time_recursive_%s", strings.ToLower(metric.Op)) field := "io_service_time_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoWaitTimeRecursive { for _, metric := range blkioStats.IoWaitTimeRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_wait_time_%s", strings.ToLower(metric.Op)) field := "io_wait_time_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }
for _, metric := range blkioStats.IoMergedRecursive { for _, metric := range blkioStats.IoMergedRecursive {
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor) device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
field := fmt.Sprintf("io_merged_recursive_%s", strings.ToLower(metric.Op)) field := "io_merged_recursive_" + strings.ToLower(metric.Op)
deviceStatMap[device][field] = metric.Value deviceStatMap[device][field] = metric.Value
} }

View File

@ -216,7 +216,7 @@ func (g *GitHub) getPullRequestFields(ctx context.Context, owner, repo string) (
return fields, err return fields, err
} }
f := fmt.Sprintf("%s_pull_requests", class) f := class + "_pull_requests"
fields[f] = searchResult.GetTotal() fields[f] = searchResult.GetTotal()
} }

View File

@ -221,7 +221,7 @@ func TestBodyAndContentEncoding(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
address := fmt.Sprintf("http://%s", ts.Listener.Addr().String()) address := "http://" + ts.Listener.Addr().String()
tests := []struct { tests := []struct {
name string name string
@ -320,7 +320,7 @@ func TestOAuthClientCredentialsGrant(t *testing.T) {
var token = "2YotnFZFEjr1zCsicMWpAA" var token = "2YotnFZFEjr1zCsicMWpAA"
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {

View File

@ -282,8 +282,8 @@ func (h *HTTPResponse) httpGather(u string) (map[string]interface{}, map[string]
h.setBodyReadError("The body of the HTTP Response is too large", bodyBytes, fields, tags) h.setBodyReadError("The body of the HTTP Response is too large", bodyBytes, fields, tags)
return fields, tags, nil return fields, tags, nil
} else if err != nil { } else if err != nil {
h.setBodyReadError(fmt.Sprintf("Failed to read body of HTTP Response : %s", err.Error()), bodyBytes, fields, tags) h.setBodyReadError("Failed to read body of HTTP Response : "+err.Error(), bodyBytes, fields, tags)
return fields, tags, nil return fields, tags, nil //nolint:nilerr // error is handled properly
} }
// Add the body of the response if expected // Add the body of the response if expected

View File

@ -133,7 +133,7 @@ func (i *Icinga2) gatherObjects(acc telegraf.Accumulator, checks ResultObject, o
"port": serverURL.Port(), "port": serverURL.Port(),
} }
acc.AddFields(fmt.Sprintf("icinga2_%s", objectType), fields, tags) acc.AddFields("icinga2_"+objectType, fields, tags)
} }
} }

View File

@ -343,7 +343,7 @@ func (h *InfluxDBListener) handleWriteInternalParser(res http.ResponseWriter, re
case 1: case 1:
partialErrorString = firstParseErrorStr partialErrorString = firstParseErrorStr
case 2: case 2:
partialErrorString = fmt.Sprintf("%s (and 1 other parse error)", firstParseErrorStr) partialErrorString = firstParseErrorStr + " (and 1 other parse error)"
default: default:
partialErrorString = fmt.Sprintf("%s (and %d other parse errors)", firstParseErrorStr, parseErrorCount-1) partialErrorString = fmt.Sprintf("%s (and %d other parse errors)", firstParseErrorStr, parseErrorCount-1)
} }
@ -459,7 +459,7 @@ func (h *InfluxDBListener) handleWriteUpstreamParser(res http.ResponseWriter, re
case 1: case 1:
partialErrorString = firstParseErrorStr partialErrorString = firstParseErrorStr
case 2: case 2:
partialErrorString = fmt.Sprintf("%s (and 1 other parse error)", firstParseErrorStr) partialErrorString = firstParseErrorStr + " (and 1 other parse error)"
default: default:
partialErrorString = fmt.Sprintf("%s (and %d other parse errors)", firstParseErrorStr, parseErrorCount-1) partialErrorString = fmt.Sprintf("%s (and %d other parse errors)", firstParseErrorStr, parseErrorCount-1)
} }

View File

@ -4,7 +4,6 @@ import (
"bytes" "bytes"
"crypto/tls" "crypto/tls"
"crypto/x509" "crypto/x509"
"fmt"
"io" "io"
"net/http" "net/http"
"net/url" "net/url"
@ -266,7 +265,7 @@ func TestWriteKeepDatabase(t *testing.T) {
testMsgWithDB := "cpu_load_short,host=server01,database=wrongdb value=12.0 1422568543702900257\n" testMsgWithDB := "cpu_load_short,host=server01,database=wrongdb value=12.0 1422568543702900257\n"
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
listener.DatabaseTag = "database" listener.DatabaseTag = "database"
@ -353,7 +352,7 @@ func TestWriteRetentionPolicyTag(t *testing.T) {
// http listener should add a newline at the end of the buffer if it's not there // http listener should add a newline at the end of the buffer if it's not there
func TestWriteNoNewline(t *testing.T) { func TestWriteNoNewline(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -379,7 +378,7 @@ func TestWriteNoNewline(t *testing.T) {
func TestPartialWrite(t *testing.T) { func TestPartialWrite(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -413,7 +412,7 @@ func TestWriteMaxLineSizeIncrease(t *testing.T) {
require.NoError(t, err) require.NoError(t, err)
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := &InfluxDBListener{ listener := &InfluxDBListener{
Log: testutil.Logger{}, Log: testutil.Logger{},
ServiceAddress: "localhost:0", ServiceAddress: "localhost:0",
@ -441,7 +440,7 @@ func TestWriteVerySmallMaxBody(t *testing.T) {
require.NoError(t, err) require.NoError(t, err)
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := &InfluxDBListener{ listener := &InfluxDBListener{
Log: testutil.Logger{}, Log: testutil.Logger{},
ServiceAddress: "localhost:0", ServiceAddress: "localhost:0",
@ -474,7 +473,7 @@ func TestWriteLargeLine(t *testing.T) {
hugeMetricString := string(hugeMetric) hugeMetricString := string(hugeMetric)
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := &InfluxDBListener{ listener := &InfluxDBListener{
Log: testutil.Logger{}, Log: testutil.Logger{},
ServiceAddress: "localhost:0", ServiceAddress: "localhost:0",
@ -554,7 +553,7 @@ func TestWriteLargeLine(t *testing.T) {
// test that writing gzipped data works // test that writing gzipped data works
func TestWriteGzippedData(t *testing.T) { func TestWriteGzippedData(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -635,7 +634,7 @@ func TestWriteHighTraffic(t *testing.T) {
func TestReceive404ForInvalidEndpoint(t *testing.T) { func TestReceive404ForInvalidEndpoint(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -655,7 +654,7 @@ func TestReceive404ForInvalidEndpoint(t *testing.T) {
func TestWriteInvalid(t *testing.T) { func TestWriteInvalid(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -675,7 +674,7 @@ func TestWriteInvalid(t *testing.T) {
func TestWriteEmpty(t *testing.T) { func TestWriteEmpty(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -695,7 +694,7 @@ func TestWriteEmpty(t *testing.T) {
func TestQuery(t *testing.T) { func TestQuery(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -716,7 +715,7 @@ func TestQuery(t *testing.T) {
func TestPing(t *testing.T) { func TestPing(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -738,7 +737,7 @@ func TestPing(t *testing.T) {
func TestPingVerbose(t *testing.T) { func TestPingVerbose(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -760,7 +759,7 @@ func TestPingVerbose(t *testing.T) {
func TestWriteWithPrecision(t *testing.T) { func TestWriteWithPrecision(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -787,7 +786,7 @@ func TestWriteWithPrecision(t *testing.T) {
func TestWriteWithPrecisionNoTimestamp(t *testing.T) { func TestWriteWithPrecisionNoTimestamp(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
listener.timeFunc = func() time.Time { listener.timeFunc = func() time.Time {

View File

@ -92,7 +92,7 @@ func (h *InfluxDBV2Listener) Gather(_ telegraf.Accumulator) error {
func (h *InfluxDBV2Listener) routes() { func (h *InfluxDBV2Listener) routes() {
credentials := "" credentials := ""
if h.Token != "" { if h.Token != "" {
credentials = fmt.Sprintf("Token %s", h.Token) credentials = "Token " + h.Token
} }
authHandler := internal.GenericAuthHandler(credentials, authHandler := internal.GenericAuthHandler(credentials,
func(_ http.ResponseWriter) { func(_ http.ResponseWriter) {

View File

@ -4,7 +4,6 @@ import (
"bytes" "bytes"
"crypto/tls" "crypto/tls"
"crypto/x509" "crypto/x509"
"fmt"
"io" "io"
"net/http" "net/http"
"net/url" "net/url"
@ -154,7 +153,7 @@ func TestWriteTokenAuth(t *testing.T) {
req, err := http.NewRequest("POST", createURL(listener, "http", "/api/v2/write", "bucket=mybucket"), bytes.NewBuffer([]byte(testMsg))) req, err := http.NewRequest("POST", createURL(listener, "http", "/api/v2/write", "bucket=mybucket"), bytes.NewBuffer([]byte(testMsg)))
require.NoError(t, err) require.NoError(t, err)
req.Header.Set("Authorization", fmt.Sprintf("Token %s", token)) req.Header.Set("Authorization", "Token "+token)
resp, err := client.Do(req) resp, err := client.Do(req)
require.NoError(t, err) require.NoError(t, err)
require.NoError(t, resp.Body.Close()) require.NoError(t, resp.Body.Close())
@ -216,7 +215,7 @@ func TestWriteKeepBucket(t *testing.T) {
// http listener should add a newline at the end of the buffer if it's not there // http listener should add a newline at the end of the buffer if it's not there
func TestWriteNoNewline(t *testing.T) { func TestWriteNoNewline(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -242,7 +241,7 @@ func TestWriteNoNewline(t *testing.T) {
func TestAllOrNothing(t *testing.T) { func TestAllOrNothing(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -265,7 +264,7 @@ func TestWriteMaxLineSizeIncrease(t *testing.T) {
require.NoError(t, err) require.NoError(t, err)
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := &InfluxDBV2Listener{ listener := &InfluxDBV2Listener{
Log: testutil.Logger{}, Log: testutil.Logger{},
ServiceAddress: "localhost:0", ServiceAddress: "localhost:0",
@ -293,7 +292,7 @@ func TestWriteVerySmallMaxBody(t *testing.T) {
require.NoError(t, err) require.NoError(t, err)
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := &InfluxDBV2Listener{ listener := &InfluxDBV2Listener{
Log: testutil.Logger{}, Log: testutil.Logger{},
ServiceAddress: "localhost:0", ServiceAddress: "localhost:0",
@ -469,7 +468,7 @@ func TestWriteHighTraffic(t *testing.T) {
func TestReceive404ForInvalidEndpoint(t *testing.T) { func TestReceive404ForInvalidEndpoint(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -489,7 +488,7 @@ func TestReceive404ForInvalidEndpoint(t *testing.T) {
func TestWriteInvalid(t *testing.T) { func TestWriteInvalid(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -509,7 +508,7 @@ func TestWriteInvalid(t *testing.T) {
func TestWriteEmpty(t *testing.T) { func TestWriteEmpty(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser
@ -550,7 +549,7 @@ func TestReady(t *testing.T) {
func TestWriteWithPrecision(t *testing.T) { func TestWriteWithPrecision(t *testing.T) {
for _, tc := range parserTestCases { for _, tc := range parserTestCases {
t.Run(fmt.Sprintf("parser %s", tc.parser), func(t *testing.T) { t.Run("parser "+tc.parser, func(t *testing.T) {
listener := newTestListener() listener := newTestListener()
listener.ParserType = tc.parser listener.ParserType = tc.parser

View File

@ -4,7 +4,6 @@ package intel_baseband
import ( import (
"errors" "errors"
"fmt"
"testing" "testing"
"time" "time"
@ -244,7 +243,7 @@ func TestParseOperationName(t *testing.T) {
logConnector := prepareLogConnMock() logConnector := prepareLogConnMock()
require.NotNil(t, logConnector) require.NotNil(t, logConnector)
for _, tc := range testCases { for _, tc := range testCases {
t.Run(fmt.Sprintf("expected %s", tc.expected), func(t *testing.T) { t.Run("expected "+tc.expected, func(t *testing.T) {
operationName := logConnector.parseOperationName(tc.input) operationName := logConnector.parseOperationName(tc.input)
require.Equal(t, tc.expected, operationName) require.Equal(t, tc.expected, operationName)
}) })

View File

@ -2,7 +2,6 @@ package kube_inventory
import ( import (
"context" "context"
"fmt"
"strings" "strings"
corev1 "k8s.io/api/core/v1" corev1 "k8s.io/api/core/v1"
@ -16,7 +15,7 @@ func collectPods(ctx context.Context, acc telegraf.Accumulator, ki *KubernetesIn
var err error var err error
if ki.KubeletURL != "" { if ki.KubeletURL != "" {
err = ki.queryPodsFromKubelet(fmt.Sprintf("%s/pods", ki.KubeletURL), listRef) err = ki.queryPodsFromKubelet(ki.KubeletURL+"/pods", listRef)
} else { } else {
listRef, err = ki.client.getPods(ctx, ki.NodeName) listRef, err = ki.client.getPods(ctx, ki.NodeName)
} }

View File

@ -159,7 +159,7 @@ func getNodeAddress(addresses []v1.NodeAddress) string {
func (k *Kubernetes) gatherSummary(baseURL string, acc telegraf.Accumulator) error { func (k *Kubernetes) gatherSummary(baseURL string, acc telegraf.Accumulator) error {
summaryMetrics := &SummaryMetrics{} summaryMetrics := &SummaryMetrics{}
err := k.LoadJSON(fmt.Sprintf("%s/stats/summary", baseURL), summaryMetrics) err := k.LoadJSON(baseURL+"/stats/summary", summaryMetrics)
if err != nil { if err != nil {
return err return err
} }
@ -224,7 +224,7 @@ func buildNodeMetrics(summaryMetrics *SummaryMetrics, acc telegraf.Accumulator)
func (k *Kubernetes) gatherPodInfo(baseURL string) ([]Item, error) { func (k *Kubernetes) gatherPodInfo(baseURL string) ([]Item, error) {
var podAPI Pods var podAPI Pods
err := k.LoadJSON(fmt.Sprintf("%s/pods", baseURL), &podAPI) err := k.LoadJSON(baseURL+"/pods", &podAPI)
if err != nil { if err != nil {
return nil, err return nil, err
} }

View File

@ -284,7 +284,7 @@ func (logstash *Logstash) gatherPluginsStats(
if strings.HasPrefix(k, "bulk_requests") { if strings.HasPrefix(k, "bulk_requests") {
continue continue
} }
newKey := fmt.Sprintf("bulk_requests_%s", k) newKey := "bulk_requests_" + k
flattener.Fields[newKey] = v flattener.Fields[newKey] = v
delete(flattener.Fields, k) delete(flattener.Fields, k)
} }
@ -307,7 +307,7 @@ func (logstash *Logstash) gatherPluginsStats(
if strings.HasPrefix(k, "documents") { if strings.HasPrefix(k, "documents") {
continue continue
} }
newKey := fmt.Sprintf("documents_%s", k) newKey := "documents_" + k
flattener.Fields[newKey] = v flattener.Fields[newKey] = v
delete(flattener.Fields, k) delete(flattener.Fields, k)
} }

View File

@ -58,7 +58,7 @@ func (p *ReportsParams) String() string {
func NewChimpAPI(apiKey string, log telegraf.Logger) *ChimpAPI { func NewChimpAPI(apiKey string, log telegraf.Logger) *ChimpAPI {
u := &url.URL{} u := &url.URL{}
u.Scheme = "https" u.Scheme = "https"
u.Host = fmt.Sprintf("%s.api.mailchimp.com", mailchimpDatacenter.FindString(apiKey)) u.Host = mailchimpDatacenter.FindString(apiKey) + ".api.mailchimp.com"
u.User = url.UserPassword("", apiKey) u.User = url.UserPassword("", apiKey)
return &ChimpAPI{url: u, log: log} return &ChimpAPI{url: u, log: log}
} }

View File

@ -213,7 +213,7 @@ func (m *Monit) Init() error {
} }
func (m *Monit) Gather(acc telegraf.Accumulator) error { func (m *Monit) Gather(acc telegraf.Accumulator) error {
req, err := http.NewRequest("GET", fmt.Sprintf("%s/_status?format=xml", m.Address), nil) req, err := http.NewRequest("GET", m.Address+"/_status?format=xml", nil)
if err != nil { if err != nil {
return err return err
} }

View File

@ -137,7 +137,7 @@ func (n *NeptuneApex) parseXML(acc telegraf.Accumulator, data []byte) error {
} }
// Find Amp and Watt probes and add them as fields. // Find Amp and Watt probes and add them as fields.
// Remove the redundant probe. // Remove the redundant probe.
if pos := findProbe(fmt.Sprintf("%sW", o.Name), r.Probe); pos > -1 { if pos := findProbe(o.Name+"W", r.Probe); pos > -1 {
value, err := strconv.ParseFloat( value, err := strconv.ParseFloat(
strings.TrimSpace(r.Probe[pos].Value), 64) strings.TrimSpace(r.Probe[pos].Value), 64)
if err != nil { if err != nil {
@ -149,7 +149,7 @@ func (n *NeptuneApex) parseXML(acc telegraf.Accumulator, data []byte) error {
r.Probe[pos] = r.Probe[len(r.Probe)-1] r.Probe[pos] = r.Probe[len(r.Probe)-1]
r.Probe = r.Probe[:len(r.Probe)-1] r.Probe = r.Probe[:len(r.Probe)-1]
} }
if pos := findProbe(fmt.Sprintf("%sA", o.Name), r.Probe); pos > -1 { if pos := findProbe(o.Name+"A", r.Probe); pos > -1 {
value, err := strconv.ParseFloat( value, err := strconv.ParseFloat(
strings.TrimSpace(r.Probe[pos].Value), 64) strings.TrimSpace(r.Probe[pos].Value), 64)
if err != nil { if err != nil {
@ -245,7 +245,7 @@ func parseTime(val string, tz float64) (time.Time, error) {
} }
func (n *NeptuneApex) sendRequest(server string) ([]byte, error) { func (n *NeptuneApex) sendRequest(server string) ([]byte, error) {
url := fmt.Sprintf("%s/cgi-bin/status.xml", server) url := server + "/cgi-bin/status.xml"
resp, err := n.httpClient.Get(url) resp, err := n.httpClient.Get(url)
if err != nil { if err != nil {
return nil, fmt.Errorf("http GET failed: %w", err) return nil, fmt.Errorf("http GET failed: %w", err)

View File

@ -55,11 +55,11 @@ func TestNginxGeneratesMetrics(t *testing.T) {
defer ts.Close() defer ts.Close()
n := &Nginx{ n := &Nginx{
Urls: []string{fmt.Sprintf("%s/stub_status", ts.URL)}, Urls: []string{ts.URL + "/stub_status"},
} }
nt := &Nginx{ nt := &Nginx{
Urls: []string{fmt.Sprintf("%s/tengine_status", ts.URL)}, Urls: []string{ts.URL + "/tengine_status"},
} }
var accNginx testutil.Accumulator var accNginx testutil.Accumulator

View File

@ -264,7 +264,7 @@ func TestNginxPlusGeneratesMetrics(t *testing.T) {
defer ts.Close() defer ts.Close()
n := &NginxPlus{ n := &NginxPlus{
Urls: []string{fmt.Sprintf("%s/status", ts.URL)}, Urls: []string{ts.URL + "/status"},
} }
var acc testutil.Accumulator var acc testutil.Accumulator

View File

@ -8,8 +8,9 @@ import (
"net/url" "net/url"
"testing" "testing"
"github.com/influxdata/telegraf/testutil"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
"github.com/influxdata/telegraf/testutil"
) )
const processesPayload = ` const processesPayload = `
@ -1532,7 +1533,7 @@ func TestUnknownContentType(t *testing.T) {
func prepareAddr(t *testing.T, ts *httptest.Server) (*url.URL, string, string) { func prepareAddr(t *testing.T, ts *httptest.Server) (*url.URL, string, string) {
t.Helper() t.Helper()
addr, err := url.Parse(fmt.Sprintf("%s/api", ts.URL)) addr, err := url.Parse(ts.URL + "/api")
require.NoError(t, err) require.NoError(t, err)
host, port, err := net.SplitHostPort(addr.Host) host, port, err := net.SplitHostPort(addr.Host)
@ -1561,7 +1562,7 @@ func prepareEndpoint(t *testing.T, path string, payload string) (*httptest.Serve
})) }))
n := &NginxPlusAPI{ n := &NginxPlusAPI{
Urls: []string{fmt.Sprintf("%s/api", ts.URL)}, Urls: []string{ts.URL + "/api"},
APIVersion: defaultAPIVersion, APIVersion: defaultAPIVersion,
} }

View File

@ -178,7 +178,7 @@ func TestNginxPlusGeneratesMetrics(t *testing.T) {
defer ts.Close() defer ts.Close()
n := &NginxSTS{ n := &NginxSTS{
Urls: []string{fmt.Sprintf("%s/status", ts.URL)}, Urls: []string{ts.URL + "/status"},
} }
var acc testutil.Accumulator var acc testutil.Accumulator

View File

@ -57,7 +57,7 @@ func TestNginxUpstreamCheckData(test *testing.T) {
defer testServer.Close() defer testServer.Close()
check := NewNginxUpstreamCheck() check := NewNginxUpstreamCheck()
check.URL = fmt.Sprintf("%s/status", testServer.URL) check.URL = testServer.URL + "/status"
var accumulator testutil.Accumulator var accumulator testutil.Accumulator
@ -68,34 +68,34 @@ func TestNginxUpstreamCheckData(test *testing.T) {
test, test,
"nginx_upstream_check", "nginx_upstream_check",
map[string]interface{}{ map[string]interface{}{
"status": string("up"), "status": "up",
"status_code": uint8(1), "status_code": uint8(1),
"rise": uint64(1000), "rise": uint64(1000),
"fall": uint64(0), "fall": uint64(0),
}, },
map[string]string{ map[string]string{
"upstream": string("upstream-1"), "upstream": "upstream-1",
"type": string("http"), "type": "http",
"name": string("127.0.0.1:8081"), "name": "127.0.0.1:8081",
"port": string("0"), "port": "0",
"url": fmt.Sprintf("%s/status", testServer.URL), "url": testServer.URL + "/status",
}) })
accumulator.AssertContainsTaggedFields( accumulator.AssertContainsTaggedFields(
test, test,
"nginx_upstream_check", "nginx_upstream_check",
map[string]interface{}{ map[string]interface{}{
"status": string("down"), "status": "down",
"status_code": uint8(2), "status_code": uint8(2),
"rise": uint64(0), "rise": uint64(0),
"fall": uint64(2000), "fall": uint64(2000),
}, },
map[string]string{ map[string]string{
"upstream": string("upstream-2"), "upstream": "upstream-2",
"type": string("tcp"), "type": "tcp",
"name": string("127.0.0.1:8082"), "name": "127.0.0.1:8082",
"port": string("8080"), "port": "8080",
"url": fmt.Sprintf("%s/status", testServer.URL), "url": testServer.URL + "/status",
}) })
} }
@ -119,7 +119,7 @@ func TestNginxUpstreamCheckRequest(test *testing.T) {
defer testServer.Close() defer testServer.Close()
check := NewNginxUpstreamCheck() check := NewNginxUpstreamCheck()
check.URL = fmt.Sprintf("%s/status", testServer.URL) check.URL = testServer.URL + "/status"
check.Headers["X-test"] = "test-value" check.Headers["X-test"] = "test-value"
check.HostHeader = "status.local" check.HostHeader = "status.local"
check.Username = "user" check.Username = "user"

View File

@ -215,7 +215,7 @@ func TestNginxPlusGeneratesMetrics(t *testing.T) {
defer ts.Close() defer ts.Close()
n := &NginxVTS{ n := &NginxVTS{
Urls: []string{fmt.Sprintf("%s/status", ts.URL)}, Urls: []string{ts.URL + "/status"},
} }
var acc testutil.Accumulator var acc testutil.Accumulator

View File

@ -21,7 +21,7 @@ import (
// The `data` field contains a list of commands to execute with // The `data` field contains a list of commands to execute with
// the \n character after every command. // the \n character after every command.
func (p *PowerdnsRecursor) gatherFromV1Server(address string, acc telegraf.Accumulator) error { func (p *PowerdnsRecursor) gatherFromV1Server(address string, acc telegraf.Accumulator) error {
recvSocket := filepath.Join(p.SocketDir, fmt.Sprintf("pdns_recursor_telegraf%s", uuid.New().String())) recvSocket := filepath.Join(p.SocketDir, "pdns_recursor_telegraf"+uuid.New().String())
laddr, err := net.ResolveUnixAddr("unixgram", recvSocket) laddr, err := net.ResolveUnixAddr("unixgram", recvSocket)
if err != nil { if err != nil {

View File

@ -19,7 +19,7 @@ import (
// Datagram 1 => status: uint32 // Datagram 1 => status: uint32
// Datagram 2 => data: byte[] (max 16_384 bytes) // Datagram 2 => data: byte[] (max 16_384 bytes)
func (p *PowerdnsRecursor) gatherFromV2Server(address string, acc telegraf.Accumulator) error { func (p *PowerdnsRecursor) gatherFromV2Server(address string, acc telegraf.Accumulator) error {
recvSocket := filepath.Join(p.SocketDir, fmt.Sprintf("pdns_recursor_telegraf%s", uuid.New().String())) recvSocket := filepath.Join(p.SocketDir, "pdns_recursor_telegraf"+uuid.New().String())
laddr, err := net.ResolveUnixAddr("unixgram", recvSocket) laddr, err := net.ResolveUnixAddr("unixgram", recvSocket)
if err != nil { if err != nil {

View File

@ -366,7 +366,7 @@ func (p *Procstat) supervisorPIDs() ([]string, map[string]map[string]string, err
func (p *Procstat) systemdUnitPIDs() ([]PidsTags, error) { func (p *Procstat) systemdUnitPIDs() ([]PidsTags, error) {
if p.IncludeSystemdChildren { if p.IncludeSystemdChildren {
p.CGroup = fmt.Sprintf("systemd/system.slice/%s", p.SystemdUnit) p.CGroup = "systemd/system.slice/" + p.SystemdUnit
return p.cgroupPIDs() return p.cgroupPIDs()
} }

View File

@ -58,7 +58,7 @@ func TestRaindropsGeneratesMetrics(t *testing.T) {
defer ts.Close() defer ts.Close()
n := &Raindrops{ n := &Raindrops{
Urls: []string{fmt.Sprintf("%s/_raindrops", ts.URL)}, Urls: []string{ts.URL + "/_raindrops"},
httpClient: &http.Client{Transport: &http.Transport{ httpClient: &http.Client{Transport: &http.Transport{
ResponseHeaderTimeout: 3 * time.Second, ResponseHeaderTimeout: 3 * time.Second,
}}, }},

View File

@ -3,7 +3,6 @@
package ras package ras
import ( import (
"fmt"
"testing" "testing"
"github.com/stretchr/testify/require" "github.com/stretchr/testify/require"
@ -23,14 +22,14 @@ func TestUpdateCounters(t *testing.T) {
if metric == processorBase { if metric == processorBase {
// processor_base_errors is sum of other seven errors: internal_timer_errors, smm_handler_code_access_violation_errors, // processor_base_errors is sum of other seven errors: internal_timer_errors, smm_handler_code_access_violation_errors,
// internal_parity_errors, frc_errors, external_mce_errors, microcode_rom_parity_errors and unclassified_mce_errors // internal_parity_errors, frc_errors, external_mce_errors, microcode_rom_parity_errors and unclassified_mce_errors
require.Equal(t, int64(7), value, fmt.Sprintf("%s should have value of 7", processorBase)) require.Equal(t, int64(7), value, processorBase+" should have value of 7")
} else { } else {
require.Equal(t, int64(1), value, fmt.Sprintf("%s should have value of 1", metric)) require.Equal(t, int64(1), value, metric+" should have value of 1")
} }
} }
for metric, value := range ras.serverCounters { for metric, value := range ras.serverCounters {
require.Equal(t, int64(1), value, fmt.Sprintf("%s should have value of 1", metric)) require.Equal(t, int64(1), value, metric+" should have value of 1")
} }
} }
@ -102,9 +101,9 @@ func TestMultipleSockets(t *testing.T) {
for _, metricData := range ras.cpuSocketCounters { for _, metricData := range ras.cpuSocketCounters {
for metric, value := range metricData { for metric, value := range metricData {
if metric == levelTwoCache { if metric == levelTwoCache {
require.Equal(t, int64(1), value, fmt.Sprintf("%s should have value of 1", levelTwoCache)) require.Equal(t, int64(1), value, levelTwoCache+" should have value of 1")
} else { } else {
require.Equal(t, int64(0), value, fmt.Sprintf("%s should have value of 0", metric)) require.Equal(t, int64(0), value, metric+" should have value of 0")
} }
} }
} }
@ -125,11 +124,11 @@ func TestEmptyDatabase(t *testing.T) {
require.Len(t, ras.serverCounters, 2, "Should contain default counters for server") require.Len(t, ras.serverCounters, 2, "Should contain default counters for server")
for metric, value := range ras.cpuSocketCounters[0] { for metric, value := range ras.cpuSocketCounters[0] {
require.Equal(t, int64(0), value, fmt.Sprintf("%s should have value of 0", metric)) require.Equal(t, int64(0), value, metric+" should have value of 0")
} }
for metric, value := range ras.serverCounters { for metric, value := range ras.serverCounters {
require.Equal(t, int64(0), value, fmt.Sprintf("%s should have value of 0", metric)) require.Equal(t, int64(0), value, metric+" should have value of 0")
} }
} }

View File

@ -693,7 +693,7 @@ func coerceType(value interface{}, typ reflect.Type) reflect.Value {
value = float64(0) value = float64(0)
} }
default: default:
panic(fmt.Sprintf("unhandled destination type %s", typ.Kind().String())) panic("unhandled destination type " + typ.Kind().String())
} }
case int, int8, int16, int32, int64: case int, int8, int16, int32, int64:
switch typ.Kind() { switch typ.Kind() {
@ -704,7 +704,7 @@ func coerceType(value interface{}, typ reflect.Type) reflect.Value {
case reflect.Float64: case reflect.Float64:
value = float64(reflect.ValueOf(sourceType).Int()) value = float64(reflect.ValueOf(sourceType).Int())
default: default:
panic(fmt.Sprintf("unhandled destination type %s", typ.Kind().String())) panic("unhandled destination type " + typ.Kind().String())
} }
case uint, uint8, uint16, uint32, uint64: case uint, uint8, uint16, uint32, uint64:
switch typ.Kind() { switch typ.Kind() {
@ -715,7 +715,7 @@ func coerceType(value interface{}, typ reflect.Type) reflect.Value {
case reflect.Float64: case reflect.Float64:
value = float64(reflect.ValueOf(sourceType).Uint()) value = float64(reflect.ValueOf(sourceType).Uint())
default: default:
panic(fmt.Sprintf("unhandled destination type %s", typ.Kind().String())) panic("unhandled destination type " + typ.Kind().String())
} }
case float32, float64: case float32, float64:
switch typ.Kind() { switch typ.Kind() {
@ -726,7 +726,7 @@ func coerceType(value interface{}, typ reflect.Type) reflect.Value {
case reflect.Float64: case reflect.Float64:
// types match // types match
default: default:
panic(fmt.Sprintf("unhandled destination type %s", typ.Kind().String())) panic("unhandled destination type " + typ.Kind().String())
} }
case string: case string:
switch typ.Kind() { switch typ.Kind() {
@ -737,7 +737,7 @@ func coerceType(value interface{}, typ reflect.Type) reflect.Value {
case reflect.Float64: case reflect.Float64:
value, _ = strconv.ParseFloat(value.(string), 64) value, _ = strconv.ParseFloat(value.(string), 64)
default: default:
panic(fmt.Sprintf("unhandled destination type %s", typ.Kind().String())) panic("unhandled destination type " + typ.Kind().String())
} }
default: default:
panic(fmt.Sprintf("unhandled source type %T", sourceType)) panic(fmt.Sprintf("unhandled source type %T", sourceType))

View File

@ -996,7 +996,7 @@ func parseWearLeveling(acc telegraf.Accumulator, fields map[string]interface{},
values := []int64{min, max, avg} values := []int64{min, max, avg}
for i, submetricName := range []string{"Min", "Max", "Avg"} { for i, submetricName := range []string{"Min", "Max", "Avg"} {
fields["raw_value"] = values[i] fields["raw_value"] = values[i]
tags["name"] = fmt.Sprintf("Wear_Leveling_%s", submetricName) tags["name"] = "Wear_Leveling_" + submetricName
acc.AddFields("smart_attribute", fields, tags) acc.AddFields("smart_attribute", fields, tags)
} }

View File

@ -283,7 +283,7 @@ func (s *Stackdriver) newListTimeSeriesFilter(metricType string) string {
resourceLabelsFilter = append(resourceLabelsFilter, fmt.Sprintf(valueFmt, resourceLabel.Key, resourceLabel.Value)) resourceLabelsFilter = append(resourceLabelsFilter, fmt.Sprintf(valueFmt, resourceLabel.Key, resourceLabel.Value))
} }
if len(resourceLabelsFilter) == 1 { if len(resourceLabelsFilter) == 1 {
filterString += fmt.Sprintf(" AND %s", resourceLabelsFilter[0]) filterString += " AND " + resourceLabelsFilter[0]
} else { } else {
filterString += fmt.Sprintf(" AND (%s)", strings.Join(resourceLabelsFilter, " OR ")) filterString += fmt.Sprintf(" AND (%s)", strings.Join(resourceLabelsFilter, " OR "))
} }
@ -301,7 +301,7 @@ func (s *Stackdriver) newListTimeSeriesFilter(metricType string) string {
metricLabelsFilter = append(metricLabelsFilter, fmt.Sprintf(valueFmt, metricLabel.Key, metricLabel.Value)) metricLabelsFilter = append(metricLabelsFilter, fmt.Sprintf(valueFmt, metricLabel.Key, metricLabel.Value))
} }
if len(metricLabelsFilter) == 1 { if len(metricLabelsFilter) == 1 {
filterString += fmt.Sprintf(" AND %s", metricLabelsFilter[0]) filterString += " AND " + metricLabelsFilter[0]
} else { } else {
filterString += fmt.Sprintf(" AND (%s)", strings.Join(metricLabelsFilter, " OR ")) filterString += fmt.Sprintf(" AND (%s)", strings.Join(metricLabelsFilter, " OR "))
} }
@ -319,7 +319,7 @@ func (s *Stackdriver) newListTimeSeriesFilter(metricType string) string {
userLabelsFilter = append(userLabelsFilter, fmt.Sprintf(valueFmt, metricLabel.Key, metricLabel.Value)) userLabelsFilter = append(userLabelsFilter, fmt.Sprintf(valueFmt, metricLabel.Key, metricLabel.Value))
} }
if len(userLabelsFilter) == 1 { if len(userLabelsFilter) == 1 {
filterString += fmt.Sprintf(" AND %s", userLabelsFilter[0]) filterString += " AND " + userLabelsFilter[0]
} else { } else {
filterString += fmt.Sprintf(" AND (%s)", strings.Join(userLabelsFilter, " OR ")) filterString += fmt.Sprintf(" AND (%s)", strings.Join(userLabelsFilter, " OR "))
} }
@ -337,7 +337,7 @@ func (s *Stackdriver) newListTimeSeriesFilter(metricType string) string {
systemLabelsFilter = append(systemLabelsFilter, fmt.Sprintf(valueFmt, metricLabel.Key, metricLabel.Value)) systemLabelsFilter = append(systemLabelsFilter, fmt.Sprintf(valueFmt, metricLabel.Key, metricLabel.Value))
} }
if len(systemLabelsFilter) == 1 { if len(systemLabelsFilter) == 1 {
filterString += fmt.Sprintf(" AND %s", systemLabelsFilter[0]) filterString += " AND " + systemLabelsFilter[0]
} else { } else {
filterString += fmt.Sprintf(" AND (%s)", strings.Join(systemLabelsFilter, " OR ")) filterString += fmt.Sprintf(" AND (%s)", strings.Join(systemLabelsFilter, " OR "))
} }
@ -357,7 +357,7 @@ func (s *Stackdriver) newTimeSeriesConf(
StartTime: &timestamppb.Timestamp{Seconds: startTime.Unix()}, StartTime: &timestamppb.Timestamp{Seconds: startTime.Unix()},
} }
tsReq := &monitoringpb.ListTimeSeriesRequest{ tsReq := &monitoringpb.ListTimeSeriesRequest{
Name: fmt.Sprintf("projects/%s", s.Project), Name: "projects/" + s.Project,
Filter: filter, Filter: filter,
Interval: interval, Interval: interval,
} }
@ -493,7 +493,7 @@ func (s *Stackdriver) generatetimeSeriesConfs(
ret := []*timeSeriesConf{} ret := []*timeSeriesConf{}
req := &monitoringpb.ListMetricDescriptorsRequest{ req := &monitoringpb.ListMetricDescriptorsRequest{
Name: fmt.Sprintf("projects/%s", s.Project), Name: "projects/" + s.Project,
} }
filters := s.newListMetricDescriptorsFilters() filters := s.newListMetricDescriptorsFilters()

View File

@ -126,13 +126,13 @@ func getShowParameters(s *SystemdUnits) *[]string {
// add the fields we're interested in to the command line // add the fields we're interested in to the command line
for property := range tagMap { for property := range tagMap {
params = append(params, fmt.Sprintf("--property=%s", property)) params = append(params, "--property="+property)
} }
for property := range valueMap { for property := range valueMap {
// If a property exists within the tagMap it was already added. Do not add it again to // If a property exists within the tagMap it was already added. Do not add it again to
// keep the command line short. // keep the command line short.
if _, exists := tagMap[property]; !exists { if _, exists := tagMap[property]; !exists {
params = append(params, fmt.Sprintf("--property=%s", property)) params = append(params, "--property="+property)
} }
} }

View File

@ -310,7 +310,7 @@ func sensorsTemperaturesOld(syspath string) ([]host.TemperatureStat, error) {
c, _ := os.ReadFile(filepath.Join(filepath.Dir(file), filename[0]+"_label")) c, _ := os.ReadFile(filepath.Join(filepath.Dir(file), filename[0]+"_label"))
if c != nil { if c != nil {
//format the label from "Core 0" to "core0_" //format the label from "Core 0" to "core0_"
label = fmt.Sprintf("%s_", strings.Join(strings.Split(strings.TrimSpace(strings.ToLower(string(c))), " "), "")) label = strings.Join(strings.Split(strings.TrimSpace(strings.ToLower(string(c))), " "), "") + "_"
} }
// Get the name of the temperature you are reading // Get the name of the temperature you are reading

View File

@ -34,7 +34,7 @@ func TestTengineGeneratesMetrics(t *testing.T) {
defer ts.Close() defer ts.Close()
n := &Tengine{ n := &Tengine{
Urls: []string{fmt.Sprintf("%s/us", ts.URL)}, Urls: []string{ts.URL + "/us"},
} }
var accTengine testutil.Accumulator var accTengine testutil.Accumulator

View File

@ -381,7 +381,7 @@ func (e *Endpoint) queryResyncSummary(ctx context.Context, vsanClient *soap.Clie
} }
vsanSystemEx := types.ManagedObjectReference{ vsanSystemEx := types.ManagedObjectReference{
Type: "VsanSystemEx", Type: "VsanSystemEx",
Value: fmt.Sprintf("vsanSystemEx-%s", strings.Split(hostRefValue, "-")[1]), Value: "vsanSystemEx-" + strings.Split(hostRefValue, "-")[1],
} }
includeSummary := true includeSummary := true

View File

@ -14,7 +14,7 @@ import (
) )
// initialize test data // initialize test data
var sysDrive = fmt.Sprintf(`%s\`, os.Getenv("SystemDrive")) // C:\ var sysDrive = os.Getenv("SystemDrive") + `\` // C:\
// include Name as a tag, FreeSpace as a field, and Purpose as a known-null class property // include Name as a tag, FreeSpace as a field, and Purpose as a known-null class property
var testQuery = Query{ var testQuery = Query{

View File

@ -29,7 +29,7 @@ type poolInfo struct {
} }
func probeVersion(kstatPath string) (metricsVersion, []string, error) { func probeVersion(kstatPath string) (metricsVersion, []string, error) {
poolsDirs, err := filepath.Glob(fmt.Sprintf("%s/*/objset-*", kstatPath)) poolsDirs, err := filepath.Glob(kstatPath + "/*/objset-*")
// From the docs: the only possible returned error is ErrBadPattern, when pattern is malformed. // From the docs: the only possible returned error is ErrBadPattern, when pattern is malformed.
// Because of this we need to determine how to fallback differently. // Because of this we need to determine how to fallback differently.
@ -42,7 +42,7 @@ func probeVersion(kstatPath string) (metricsVersion, []string, error) {
} }
// Fallback to the old kstat in case of an older ZFS version. // Fallback to the old kstat in case of an older ZFS version.
poolsDirs, err = filepath.Glob(fmt.Sprintf("%s/*/io", kstatPath)) poolsDirs, err = filepath.Glob(kstatPath + "/*/io")
if err != nil { if err != nil {
return unknown, poolsDirs, err return unknown, poolsDirs, err
} }

View File

@ -165,7 +165,7 @@ func (adx *AzureDataExplorer) pushMetrics(ctx context.Context, format ingest.Fil
length := len(metricsArray) length := len(metricsArray)
adx.Log.Debugf("Writing %d metrics to table %q", length, tableName) adx.Log.Debugf("Writing %d metrics to table %q", length, tableName)
reader := bytes.NewReader(metricsArray) reader := bytes.NewReader(metricsArray)
mapping := ingest.IngestionMappingRef(fmt.Sprintf("%s_mapping", tableName), ingest.JSON) mapping := ingest.IngestionMappingRef(tableName+"_mapping", ingest.JSON)
if metricIngestor != nil { if metricIngestor != nil {
if _, err := metricIngestor.FromReader(ctx, reader, format, mapping); err != nil { if _, err := metricIngestor.FromReader(ctx, reader, format, mapping); err != nil {
adx.Log.Errorf("sending ingestion request to Azure Data Explorer for table %q failed: %v", tableName, err) adx.Log.Errorf("sending ingestion request to Azure Data Explorer for table %q failed: %v", tableName, err)

View File

@ -4,7 +4,6 @@ import (
"context" "context"
"encoding/base64" "encoding/base64"
"errors" "errors"
"fmt"
"runtime" "runtime"
"sync" "sync"
"testing" "testing"
@ -225,7 +224,7 @@ func (r *stubResult) Get(ctx context.Context) (string, error) {
case err := <-r.err: case err := <-r.err:
return "", err return "", err
case <-r.done: case <-r.done:
return fmt.Sprintf("id-%s", r.metricIds[0]), nil return "id-" + r.metricIds[0], nil
} }
} }

View File

@ -62,7 +62,7 @@ func TestMethod(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -132,7 +132,7 @@ func TestHTTPClientConfig(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -201,7 +201,7 @@ func TestStatusCode(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -287,7 +287,7 @@ func TestContentType(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -335,7 +335,7 @@ func TestContentEncodingGzip(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -396,7 +396,7 @@ func TestBasicAuth(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -453,7 +453,7 @@ func TestOAuthClientCredentialsGrant(t *testing.T) {
var token = "2YotnFZFEjr1zCsicMWpAA" var token = "2YotnFZFEjr1zCsicMWpAA"
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -556,7 +556,7 @@ func TestOAuthAuthorizationCodeGrant(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tmpDir := t.TempDir() tmpDir := t.TempDir()
@ -654,7 +654,7 @@ func TestDefaultUserAgent(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
t.Run("default-user-agent", func(t *testing.T) { t.Run("default-user-agent", func(t *testing.T) {
@ -683,7 +683,7 @@ func TestBatchedUnbatched(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
client := &HTTP{ client := &HTTP{
@ -734,7 +734,7 @@ func TestAwsCredentials(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {

View File

@ -63,7 +63,7 @@ func TestHTTP_CreateDatabase(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
successResponse := []byte(`{"results": [{"statement_id": 0}]}`) successResponse := []byte(`{"results": [{"statement_id": 0}]}`)
@ -273,7 +273,7 @@ func TestHTTP_Write(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -765,7 +765,7 @@ func TestDBRPTags(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -1011,7 +1011,7 @@ func TestDBRPTagsCreateDatabaseNotCalledOnRetryAfterForbidden(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
handlers := &MockHandlerChain{ handlers := &MockHandlerChain{
@ -1084,7 +1084,7 @@ func TestDBRPTagsCreateDatabaseCalledOnDatabaseNotFound(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
handlers := &MockHandlerChain{ handlers := &MockHandlerChain{
@ -1176,7 +1176,7 @@ func TestDBNotFoundShouldDropMetricWhenSkipDatabaseCreateIsTrue(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
f := func(w http.ResponseWriter, r *http.Request) { f := func(w http.ResponseWriter, r *http.Request) {
switch r.URL.Path { switch r.URL.Path {

View File

@ -3,7 +3,6 @@ package loki
import ( import (
"compress/gzip" "compress/gzip"
"encoding/json" "encoding/json"
"fmt"
"io" "io"
"net/http" "net/http"
"net/http/httptest" "net/http/httptest"
@ -64,7 +63,7 @@ func TestStatusCode(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -133,7 +132,7 @@ func TestContentType(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -178,7 +177,7 @@ func TestContentEncodingGzip(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -244,7 +243,7 @@ func TestMetricNameLabel(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -294,7 +293,7 @@ func TestBasicAuth(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -341,7 +340,7 @@ func TestOAuthClientCredentialsGrant(t *testing.T) {
var token = "2YotnFZFEjr1zCsicMWpAA" var token = "2YotnFZFEjr1zCsicMWpAA"
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -408,7 +407,7 @@ func TestDefaultUserAgent(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
t.Run("default-user-agent", func(t *testing.T) { t.Run("default-user-agent", func(t *testing.T) {
@ -432,7 +431,7 @@ func TestMetricSorting(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
t.Run("out of order metrics", func(t *testing.T) { t.Run("out of order metrics", func(t *testing.T) {

View File

@ -78,7 +78,7 @@ func TestConnectAndWrite(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
testURL := fmt.Sprintf("http://%s", ts.Listener.Addr().String()) testURL := "http://" + ts.Listener.Addr().String()
testAPIKey := "a0b1c2d3-e4f5-g6h7-i8j9-k0l1m2n3o4p5" testAPIKey := "a0b1c2d3-e4f5-g6h7-i8j9-k0l1m2n3o4p5"
testCheck := "telegraf" testCheck := "telegraf"
testEntity := "entity1" testEntity := "entity1"
@ -86,7 +86,7 @@ func TestConnectAndWrite(t *testing.T) {
testHandler := "influxdb" testHandler := "influxdb"
testTagName := "myTagName" testTagName := "myTagName"
testTagValue := "myTagValue" testTagValue := "myTagValue"
expectedAuthHeader := fmt.Sprintf("Key %s", testAPIKey) expectedAuthHeader := "Key " + testAPIKey
expectedURL := fmt.Sprintf("/api/core/v2/namespaces/%s/events", testNamespace) expectedURL := fmt.Sprintf("/api/core/v2/namespaces/%s/events", testNamespace)
expectedPointName := "cpu" expectedPointName := "cpu"
expectedPointValue := float64(42) expectedPointValue := float64(42)

View File

@ -394,7 +394,7 @@ func (s *Stackdriver) sendBatch(batch []telegraf.Metric) error {
// Prepare time series request. // Prepare time series request.
timeSeriesRequest := &monitoringpb.CreateTimeSeriesRequest{ timeSeriesRequest := &monitoringpb.CreateTimeSeriesRequest{
Name: fmt.Sprintf("projects/%s", s.Project), Name: "projects/" + s.Project,
TimeSeries: timeSeries, TimeSeries: timeSeries,
} }

View File

@ -106,7 +106,7 @@ func TestWrite(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
Log: testutil.Logger{}, Log: testutil.Logger{},
client: c, client: c,
@ -134,7 +134,7 @@ func TestWriteResourceTypeAndLabels(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
ResourceType: "foo", ResourceType: "foo",
ResourceLabels: map[string]string{ ResourceLabels: map[string]string{
@ -167,7 +167,7 @@ func TestWriteTagsAsResourceLabels(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
ResourceType: "foo", ResourceType: "foo",
TagsAsResourceLabels: []string{"job_name"}, TagsAsResourceLabels: []string{"job_name"},
@ -231,7 +231,7 @@ func TestWriteMetricTypesOfficial(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
ResourceLabels: map[string]string{ ResourceLabels: map[string]string{
"mylabel": "myvalue", "mylabel": "myvalue",
@ -307,7 +307,7 @@ func TestWriteMetricTypesPath(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
ResourceLabels: map[string]string{ ResourceLabels: map[string]string{
"mylabel": "myvalue", "mylabel": "myvalue",
@ -367,7 +367,7 @@ func TestWriteAscendingTime(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
Log: testutil.Logger{}, Log: testutil.Logger{},
client: c, client: c,
@ -440,7 +440,7 @@ func TestWriteBatchable(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
Log: testutil.Logger{}, Log: testutil.Logger{},
client: c, client: c,
@ -629,7 +629,7 @@ func TestWriteIgnoredErrors(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
Log: testutil.Logger{}, Log: testutil.Logger{},
client: c, client: c,
@ -706,7 +706,7 @@ func TestGetStackdriverIntervalEndpoints(t *testing.T) {
} }
s := &Stackdriver{ s := &Stackdriver{
Project: fmt.Sprintf("projects/%s", "[PROJECT]"), Project: "projects/" + "[PROJECT]",
Namespace: "test", Namespace: "test",
Log: testutil.Logger{}, Log: testutil.Logger{},
client: c, client: c,

View File

@ -67,7 +67,7 @@ func TestMethod(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -118,7 +118,7 @@ func TestStatusCode(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
pluginFn := func() *SumoLogic { pluginFn := func() *SumoLogic {
@ -271,7 +271,7 @@ func TestContentType(t *testing.T) {
})) }))
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
plugin := tt.plugin() plugin := tt.plugin()
@ -293,7 +293,7 @@ func TestContentEncodingGzip(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
tests := []struct { tests := []struct {
@ -347,7 +347,7 @@ func TestDefaultUserAgent(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
t.Run("default-user-agent", func(t *testing.T) { t.Run("default-user-agent", func(t *testing.T) {
@ -466,7 +466,7 @@ func TestMaxRequestBodySize(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
testcases := []struct { testcases := []struct {
@ -640,7 +640,7 @@ func TestTryingToSendEmptyMetricsDoesntFail(t *testing.T) {
ts := httptest.NewServer(http.NotFoundHandler()) ts := httptest.NewServer(http.NotFoundHandler())
defer ts.Close() defer ts.Close()
u, err := url.Parse(fmt.Sprintf("http://%s", ts.Listener.Addr().String())) u, err := url.Parse("http://" + ts.Listener.Addr().String())
require.NoError(t, err) require.NoError(t, err)
metrics := make([]telegraf.Metric, 0) metrics := make([]telegraf.Metric, 0)

View File

@ -3,7 +3,6 @@ package zabbix
import ( import (
"encoding/binary" "encoding/binary"
"encoding/json" "encoding/json"
"fmt"
"net" "net"
"os" "os"
"sort" "sort"
@ -859,7 +858,7 @@ func TestBuildZabbixMetric(t *testing.T) {
1, 1,
) )
require.NoError(t, err) require.NoError(t, err)
require.Equal(t, fmt.Sprintf("%sname.value[b,bar]", keyPrefix), zm.Key) require.Equal(t, keyPrefix+"name.value[b,bar]", zm.Key)
zm, err = z.buildZabbixMetric(testutil.MustMetric( zm, err = z.buildZabbixMetric(testutil.MustMetric(
"name", "name",
@ -870,7 +869,7 @@ func TestBuildZabbixMetric(t *testing.T) {
1, 1,
) )
require.NoError(t, err) require.NoError(t, err)
require.Equal(t, fmt.Sprintf("%sname.value", keyPrefix), zm.Key) require.Equal(t, keyPrefix+"name.value", zm.Key)
} }
func TestGetHostname(t *testing.T) { func TestGetHostname(t *testing.T) {

View File

@ -59,11 +59,11 @@ func (s *Serializer) SerializeBatch(metrics []telegraf.Metric) ([]byte, error) {
switch { switch {
case strings.HasSuffix(field.Key, "_bucket"): case strings.HasSuffix(field.Key, "_bucket"):
// if bucket only, init sum, count, inf // if bucket only, init sum, count, inf
metrickeysum, promtssum := getPromTS(fmt.Sprintf("%s_sum", metricName), labels, float64(0), metric.Time()) metrickeysum, promtssum := getPromTS(metricName+"_sum", labels, float64(0), metric.Time())
if _, ok = entries[metrickeysum]; !ok { if _, ok = entries[metrickeysum]; !ok {
entries[metrickeysum] = promtssum entries[metrickeysum] = promtssum
} }
metrickeycount, promtscount := getPromTS(fmt.Sprintf("%s_count", metricName), labels, float64(0), metric.Time()) metrickeycount, promtscount := getPromTS(metricName+"_count", labels, float64(0), metric.Time())
if _, ok = entries[metrickeycount]; !ok { if _, ok = entries[metrickeycount]; !ok {
entries[metrickeycount] = promtscount entries[metrickeycount] = promtscount
} }
@ -71,7 +71,7 @@ func (s *Serializer) SerializeBatch(metrics []telegraf.Metric) ([]byte, error) {
Name: "le", Name: "le",
Value: "+Inf", Value: "+Inf",
} }
metrickeyinf, promtsinf := getPromTS(fmt.Sprintf("%s_bucket", metricName), labels, float64(0), metric.Time(), extraLabel) metrickeyinf, promtsinf := getPromTS(metricName+"_bucket", labels, float64(0), metric.Time(), extraLabel)
if _, ok = entries[metrickeyinf]; !ok { if _, ok = entries[metrickeyinf]; !ok {
entries[metrickeyinf] = promtsinf entries[metrickeyinf] = promtsinf
} }
@ -93,14 +93,14 @@ func (s *Serializer) SerializeBatch(metrics []telegraf.Metric) ([]byte, error) {
Name: "le", Name: "le",
Value: fmt.Sprint(bound), Value: fmt.Sprint(bound),
} }
metrickey, promts = getPromTS(fmt.Sprintf("%s_bucket", metricName), labels, float64(count), metric.Time(), extraLabel) metrickey, promts = getPromTS(metricName+"_bucket", labels, float64(count), metric.Time(), extraLabel)
case strings.HasSuffix(field.Key, "_sum"): case strings.HasSuffix(field.Key, "_sum"):
sum, ok := prometheus.SampleSum(field.Value) sum, ok := prometheus.SampleSum(field.Value)
if !ok { if !ok {
continue continue
} }
metrickey, promts = getPromTS(fmt.Sprintf("%s_sum", metricName), labels, sum, metric.Time()) metrickey, promts = getPromTS(metricName+"_sum", labels, sum, metric.Time())
case strings.HasSuffix(field.Key, "_count"): case strings.HasSuffix(field.Key, "_count"):
count, ok := prometheus.SampleCount(field.Value) count, ok := prometheus.SampleCount(field.Value)
if !ok { if !ok {
@ -112,12 +112,12 @@ func (s *Serializer) SerializeBatch(metrics []telegraf.Metric) ([]byte, error) {
Name: "le", Name: "le",
Value: "+Inf", Value: "+Inf",
} }
metrickeyinf, promtsinf := getPromTS(fmt.Sprintf("%s_bucket", metricName), labels, float64(count), metric.Time(), extraLabel) metrickeyinf, promtsinf := getPromTS(metricName+"_bucket", labels, float64(count), metric.Time(), extraLabel)
if minf, ok := entries[metrickeyinf]; !ok || minf.Samples[0].Value == 0 { if minf, ok := entries[metrickeyinf]; !ok || minf.Samples[0].Value == 0 {
entries[metrickeyinf] = promtsinf entries[metrickeyinf] = promtsinf
} }
metrickey, promts = getPromTS(fmt.Sprintf("%s_count", metricName), labels, float64(count), metric.Time()) metrickey, promts = getPromTS(metricName+"_count", labels, float64(count), metric.Time())
default: default:
continue continue
} }
@ -129,14 +129,14 @@ func (s *Serializer) SerializeBatch(metrics []telegraf.Metric) ([]byte, error) {
continue continue
} }
metrickey, promts = getPromTS(fmt.Sprintf("%s_sum", metricName), labels, sum, metric.Time()) metrickey, promts = getPromTS(metricName+"_sum", labels, sum, metric.Time())
case strings.HasSuffix(field.Key, "_count"): case strings.HasSuffix(field.Key, "_count"):
count, ok := prometheus.SampleCount(field.Value) count, ok := prometheus.SampleCount(field.Value)
if !ok { if !ok {
continue continue
} }
metrickey, promts = getPromTS(fmt.Sprintf("%s_count", metricName), labels, float64(count), metric.Time()) metrickey, promts = getPromTS(metricName+"_count", labels, float64(count), metric.Time())
default: default:
quantileTag, ok := metric.GetTag("quantile") quantileTag, ok := metric.GetTag("quantile")
if !ok { if !ok {

View File

@ -475,7 +475,7 @@ func (a *Accumulator) AssertDoesNotContainMeasurement(t *testing.T, measurement
defer a.Unlock() defer a.Unlock()
for _, p := range a.Metrics { for _, p := range a.Metrics {
if p.Measurement == measurement { if p.Measurement == measurement {
msg := fmt.Sprintf("found unexpected measurement %s", measurement) msg := "found unexpected measurement " + measurement
require.Fail(t, msg) require.Fail(t, msg)
} }
} }

View File

@ -137,7 +137,7 @@ func (c *Container) CheckStatus(serviceName string) error {
func (c *Container) UploadAndInstall(filename string) error { func (c *Container) UploadAndInstall(filename string) error {
basename := filepath.Base(filename) basename := filepath.Base(filename)
destination := fmt.Sprintf("/root/%s", basename) destination := "/root/" + basename
if err := c.client.Push(c.Name, filename, destination); err != nil { if err := c.client.Push(c.Name, filename, destination); err != nil {
return err return err

View File

@ -78,7 +78,7 @@ func launchTests(packageFile string, images []string) error {
fmt.Printf("starting test with %s\n", image) fmt.Printf("starting test with %s\n", image)
uuidWithHyphen := uuid.New() uuidWithHyphen := uuid.New()
name := fmt.Sprintf("telegraf-test-%s", uuidWithHyphen.String()[0:8]) name := "telegraf-test-" + uuidWithHyphen.String()[0:8]
err := runTest(image, name, packageFile) err := runTest(image, name, packageFile)
if err != nil { if err != nil {

View File

@ -147,32 +147,32 @@ func main() {
{ {
FileName: ".circleci/config.yml", FileName: ".circleci/config.yml",
Regex: `(quay\.io\/influxdb\/telegraf-ci):(\d.\d*.\d)`, Regex: `(quay\.io\/influxdb\/telegraf-ci):(\d.\d*.\d)`,
Replace: fmt.Sprintf("$1:%s", version), Replace: "$1:" + version,
}, },
{ {
FileName: ".github/workflows/govulncheck.yml", FileName: ".github/workflows/govulncheck.yml",
Regex: `(go-version-input).*`, Regex: `(go-version-input).*`,
Replace: fmt.Sprintf("$1: %s", version), Replace: "$1: " + version,
}, },
{ {
FileName: "go.mod", FileName: "go.mod",
Regex: `(go)\s(\d.\d*)`, Regex: `(go)\s(\d.\d*)`,
Replace: fmt.Sprintf("$1 %s", noPatchVersion), Replace: "$1 " + noPatchVersion,
}, },
{ {
FileName: "Makefile", FileName: "Makefile",
Regex: `(quay\.io\/influxdb\/telegraf-ci):(\d.\d*.\d)`, Regex: `(quay\.io\/influxdb\/telegraf-ci):(\d.\d*.\d)`,
Replace: fmt.Sprintf("$1:%s", version), Replace: "$1:" + version,
}, },
{ {
FileName: "README.md", FileName: "README.md",
Regex: `(Telegraf requires Go version) (\d.\d*)`, Regex: `(Telegraf requires Go version) (\d.\d*)`,
Replace: fmt.Sprintf("$1 %s", noPatchVersion), Replace: "$1 " + noPatchVersion,
}, },
{ {
FileName: "scripts/ci.docker", FileName: "scripts/ci.docker",
Regex: `(FROM golang):(\d.\d*.\d)`, Regex: `(FROM golang):(\d.\d*.\d)`,
Replace: fmt.Sprintf("$1:%s", version), Replace: "$1:" + version,
}, },
{ {
FileName: "scripts/installgo_linux.sh", FileName: "scripts/installgo_linux.sh",