chore: Fix linter findings for `revive:unused-receiver` in `plugins/inputs/[f-k]` (#16308)

This commit is contained in:
Paweł Żak 2024-12-17 15:33:30 +01:00 committed by GitHub
parent c0b3dd489e
commit 6f80899e13
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
46 changed files with 104 additions and 141 deletions

View File

@ -40,7 +40,7 @@ func (f fakeFileInfo) ModTime() time.Time { return f.modtime }
func (f fakeFileInfo) IsDir() bool { return f.isdir } func (f fakeFileInfo) IsDir() bool { return f.isdir }
func (f fakeFileInfo) Sys() interface{} { return f.sys } func (f fakeFileInfo) Sys() interface{} { return f.sys }
func (f fakeFileSystem) open(name string) (file, error) { func (fakeFileSystem) open(name string) (file, error) {
return nil, &os.PathError{Op: "Open", Path: name, Err: errors.New("not implemented by fake filesystem")} return nil, &os.PathError{Op: "Open", Path: name, Err: errors.New("not implemented by fake filesystem")}
} }

View File

@ -87,7 +87,7 @@ func (r *Fireboard) Gather(acc telegraf.Accumulator) error {
} }
// Range over all devices, gathering stats. Returns early in case of any error. // Range over all devices, gathering stats. Returns early in case of any error.
for _, s := range stats { for _, s := range stats {
r.gatherTemps(s, acc) gatherTemps(s, acc)
} }
return nil return nil
} }
@ -105,7 +105,7 @@ func scale(n int) string {
} }
// Gathers stats from a single device, adding them to the accumulator // Gathers stats from a single device, adding them to the accumulator
func (r *Fireboard) gatherTemps(s fireboardStats, acc telegraf.Accumulator) { func gatherTemps(s fireboardStats, acc telegraf.Accumulator) {
// Construct lookup for scale values // Construct lookup for scale values
for _, t := range s.LatestTemps { for _, t := range s.LatestTemps {

View File

@ -314,7 +314,7 @@ func (c *GNMI) Start(acc telegraf.Accumulator) error {
return nil return nil
} }
func (c *GNMI) Gather(_ telegraf.Accumulator) error { func (*GNMI) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -51,15 +51,15 @@ type mockServer struct {
grpcServer *grpc.Server grpcServer *grpc.Server
} }
func (s *mockServer) Capabilities(context.Context, *gnmi.CapabilityRequest) (*gnmi.CapabilityResponse, error) { func (*mockServer) Capabilities(context.Context, *gnmi.CapabilityRequest) (*gnmi.CapabilityResponse, error) {
return nil, nil return nil, nil
} }
func (s *mockServer) Get(context.Context, *gnmi.GetRequest) (*gnmi.GetResponse, error) { func (*mockServer) Get(context.Context, *gnmi.GetRequest) (*gnmi.GetResponse, error) {
return nil, nil return nil, nil
} }
func (s *mockServer) Set(context.Context, *gnmi.SetRequest) (*gnmi.SetResponse, error) { func (*mockServer) Set(context.Context, *gnmi.SetRequest) (*gnmi.SetResponse, error) {
return nil, nil return nil, nil
} }

View File

@ -89,7 +89,7 @@ func (s *tagStore) insert(subscription tagSubscription, path *pathInfo, values [
} }
} }
case "elements": case "elements":
key, match := s.getElementsKeys(path, subscription.Elements) key, match := getElementsKeys(path, subscription.Elements)
if !match || len(values) == 0 { if !match || len(values) == 0 {
return nil return nil
} }
@ -141,7 +141,7 @@ func (s *tagStore) lookup(path *pathInfo, metricTags map[string]string) map[stri
// Match elements // Match elements
for _, requiredKeys := range s.elements.required { for _, requiredKeys := range s.elements.required {
key, match := s.getElementsKeys(path, requiredKeys) key, match := getElementsKeys(path, requiredKeys)
if !match { if !match {
continue continue
} }
@ -153,7 +153,7 @@ func (s *tagStore) lookup(path *pathInfo, metricTags map[string]string) map[stri
return tags return tags
} }
func (s *tagStore) getElementsKeys(path *pathInfo, elements []string) (string, bool) { func getElementsKeys(path *pathInfo, elements []string) (string, bool) {
// Search for the required path elements and collect a ordered // Search for the required path elements and collect a ordered
// list of their values to in the form // list of their values to in the form
// elementName1={keyA=valueA,keyB=valueB,...},...,elementNameN={keyY=valueY,keyZ=valueZ} // elementName1={keyA=valueA,keyB=valueB,...},...,elementNameN={keyY=valueY,keyZ=valueZ}

View File

@ -57,7 +57,7 @@ func (gcs *GCS) Init() error {
return gcs.setOffset() return gcs.setOffset()
} }
func (gcs *GCS) SampleConfig() string { func (*GCS) SampleConfig() string {
return sampleConfig return sampleConfig
} }

View File

@ -119,10 +119,10 @@ func (c *mockHTTPClient) makeRequest(req *http.Request) (*http.Response, error)
return &resp, nil return &resp, nil
} }
func (c *mockHTTPClient) setHTTPClient(_ *http.Client) { func (*mockHTTPClient) setHTTPClient(*http.Client) {
} }
func (c *mockHTTPClient) httpClient() *http.Client { func (*mockHTTPClient) httpClient() *http.Client {
return nil return nil
} }

View File

@ -17,9 +17,7 @@ import (
"github.com/influxdata/telegraf/testutil" "github.com/influxdata/telegraf/testutil"
) )
type statServer struct{} func serverSocket(l net.Listener) {
func (s statServer) serverSocket(l net.Listener) {
for { for {
conn, err := l.Accept() conn, err := l.Accept()
if err != nil { if err != nil {
@ -151,8 +149,7 @@ func TestHaproxyGeneratesMetricsUsingSocket(t *testing.T) {
sockets[i] = sock sockets[i] = sock
defer sock.Close() //nolint:revive,gocritic // done on purpose, closing will be executed properly defer sock.Close() //nolint:revive,gocritic // done on purpose, closing will be executed properly
s := statServer{} go serverSocket(sock)
go s.serverSocket(sock)
} }
r := &HAProxy{ r := &HAProxy{
@ -191,8 +188,7 @@ func TestHaproxyGeneratesMetricsUsingTcp(t *testing.T) {
} }
defer l.Close() defer l.Close()
s := statServer{} go serverSocket(l)
go s.serverSocket(l)
r := &HAProxy{ r := &HAProxy{
Servers: []string{"tcp://" + l.Addr().String()}, Servers: []string{"tcp://" + l.Addr().String()},

View File

@ -25,7 +25,7 @@ func New() *hddtemp {
} }
// Fetch gathers disks data from hddtemp daemon. // Fetch gathers disks data from hddtemp daemon.
func (h *hddtemp) Fetch(address string) ([]Disk, error) { func (*hddtemp) Fetch(address string) ([]Disk, error) {
var ( var (
err error err error
conn net.Conn conn net.Conn

View File

@ -12,7 +12,7 @@ import (
type mockFetcher struct { type mockFetcher struct {
} }
func (h *mockFetcher) Fetch(_ string) ([]hddtemp.Disk, error) { func (*mockFetcher) Fetch(string) ([]hddtemp.Disk, error) {
return []hddtemp.Disk{ return []hddtemp.Disk{
{ {
DeviceName: "Disk1", DeviceName: "Disk1",

View File

@ -86,7 +86,7 @@ func (h *HTTP) SetParserFunc(fn telegraf.ParserFunc) {
h.parserFunc = fn h.parserFunc = fn
} }
func (h *HTTP) Start(_ telegraf.Accumulator) error { func (*HTTP) Start(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -197,7 +197,7 @@ func (h *HTTPListenerV2) Start(acc telegraf.Accumulator) error {
return nil return nil
} }
func (h *HTTPListenerV2) Gather(_ telegraf.Accumulator) error { func (*HTTPListenerV2) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -118,7 +118,7 @@ func (h *Hugepages) Gather(acc telegraf.Accumulator) error {
// gatherStatsPerNode collects root hugepages statistics // gatherStatsPerNode collects root hugepages statistics
func (h *Hugepages) gatherRootStats(acc telegraf.Accumulator) error { func (h *Hugepages) gatherRootStats(acc telegraf.Accumulator) error {
return h.gatherFromHugepagePath(acc, "hugepages_"+rootHugepages, h.rootHugepagePath, hugepagesMetricsRoot, nil) return gatherFromHugepagePath(acc, "hugepages_"+rootHugepages, h.rootHugepagePath, hugepagesMetricsRoot, nil)
} }
// gatherStatsPerNode collects hugepages statistics per NUMA node // gatherStatsPerNode collects hugepages statistics per NUMA node
@ -144,7 +144,7 @@ func (h *Hugepages) gatherStatsPerNode(acc telegraf.Accumulator) error {
"node": nodeNumber, "node": nodeNumber,
} }
hugepagesPath := filepath.Join(h.numaNodePath, nodeDir.Name(), "hugepages") hugepagesPath := filepath.Join(h.numaNodePath, nodeDir.Name(), "hugepages")
err = h.gatherFromHugepagePath(acc, "hugepages_"+perNodeHugepages, hugepagesPath, hugepagesMetricsPerNUMANode, perNodeTags) err = gatherFromHugepagePath(acc, "hugepages_"+perNodeHugepages, hugepagesPath, hugepagesMetricsPerNUMANode, perNodeTags)
if err != nil { if err != nil {
return err return err
} }
@ -152,7 +152,7 @@ func (h *Hugepages) gatherStatsPerNode(acc telegraf.Accumulator) error {
return nil return nil
} }
func (h *Hugepages) gatherFromHugepagePath(acc telegraf.Accumulator, measurement, path string, fileFilter, defaultTags map[string]string) error { func gatherFromHugepagePath(acc telegraf.Accumulator, measurement, path string, fileFilter, defaultTags map[string]string) error {
// read metrics from: hugepages/hugepages-*/* // read metrics from: hugepages/hugepages-*/*
hugepagesDirs, err := os.ReadDir(path) hugepagesDirs, err := os.ReadDir(path)
if err != nil { if err != nil {

View File

@ -121,7 +121,7 @@ func (i *Icinga2) Gather(acc telegraf.Accumulator) error {
} }
result := resultObject{} result := resultObject{}
err = i.parseObjectResponse(resp, &result) err = parseObjectResponse(resp, &result)
if err != nil { if err != nil {
return fmt.Errorf("could not parse object response: %w", err) return fmt.Errorf("could not parse object response: %w", err)
} }
@ -145,13 +145,13 @@ func (i *Icinga2) Gather(acc telegraf.Accumulator) error {
switch statusType { switch statusType {
case "ApiListener": case "ApiListener":
fields, err = i.parsePerfdataResponse(resp) fields, err = parsePerfdataResponse(resp)
case "CIB": case "CIB":
fields, err = i.parseCIBResponse(resp) fields, err = parseCIBResponse(resp)
case "IdoMysqlConnection": case "IdoMysqlConnection":
fields, err = i.parsePerfdataResponse(resp) fields, err = parsePerfdataResponse(resp)
case "IdoPgsqlConnection": case "IdoPgsqlConnection":
fields, err = i.parsePerfdataResponse(resp) fields, err = parsePerfdataResponse(resp)
} }
if err != nil { if err != nil {
@ -233,7 +233,7 @@ func (i *Icinga2) icingaRequest(address string) (*http.Response, error) {
return resp, nil return resp, nil
} }
func (i *Icinga2) parseObjectResponse(resp *http.Response, result *resultObject) error { func parseObjectResponse(resp *http.Response, result *resultObject) error {
err := json.NewDecoder(resp.Body).Decode(&result) err := json.NewDecoder(resp.Body).Decode(&result)
if err != nil { if err != nil {
return err return err
@ -246,7 +246,7 @@ func (i *Icinga2) parseObjectResponse(resp *http.Response, result *resultObject)
return nil return nil
} }
func (i *Icinga2) parseCIBResponse(resp *http.Response) (map[string]interface{}, error) { func parseCIBResponse(resp *http.Response) (map[string]interface{}, error) {
result := resultCIB{} result := resultCIB{}
err := json.NewDecoder(resp.Body).Decode(&result) err := json.NewDecoder(resp.Body).Decode(&result)
@ -262,7 +262,7 @@ func (i *Icinga2) parseCIBResponse(resp *http.Response) (map[string]interface{},
return result.Results[0].Status, nil return result.Results[0].Status, nil
} }
func (i *Icinga2) parsePerfdataResponse(resp *http.Response) (map[string]interface{}, error) { func parsePerfdataResponse(resp *http.Response) (map[string]interface{}, error) {
result := resultPerfdata{} result := resultPerfdata{}
err := json.NewDecoder(resp.Body).Decode(&result) err := json.NewDecoder(resp.Body).Decode(&result)

View File

@ -12,7 +12,7 @@ import (
) )
// Gather statistics from our infiniband cards // Gather statistics from our infiniband cards
func (i *Infiniband) Gather(acc telegraf.Accumulator) error { func (*Infiniband) Gather(acc telegraf.Accumulator) error {
rdmaDevices := rdmamap.GetRdmaDeviceList() rdmaDevices := rdmamap.GetRdmaDeviceList()
if len(rdmaDevices) == 0 { if len(rdmaDevices) == 0 {

View File

@ -76,7 +76,7 @@ func (*InfluxDBListener) SampleConfig() string {
return sampleConfig return sampleConfig
} }
func (h *InfluxDBListener) Gather(_ telegraf.Accumulator) error { func (*InfluxDBListener) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -120,7 +120,7 @@ func (h *InfluxDBV2Listener) Init() error {
return nil return nil
} }
func (h *InfluxDBV2Listener) Gather(_ telegraf.Accumulator) error { func (*InfluxDBV2Listener) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -56,7 +56,7 @@ type Baseband struct {
sockConn *socketConnector sockConn *socketConnector
} }
func (b *Baseband) SampleConfig() string { func (*Baseband) SampleConfig() string {
return sampleConfig return sampleConfig
} }

View File

@ -135,7 +135,7 @@ func (lc *logConnector) readNumVFs() error {
continue continue
} }
numVFs, err := lc.parseNumVFs(line) numVFs, err := parseNumVFs(line)
if err != nil { if err != nil {
lc.numVFs = -1 lc.numVFs = -1
return err return err
@ -189,7 +189,7 @@ func (lc *logConnector) getMetric(offsetLine int, name string) (int, *logMetric,
return offsetLine, nil, err return offsetLine, nil, err
} }
operationName := lc.parseOperationName(line) operationName := parseOperationName(line)
if len(operationName) == 0 { if len(operationName) == 0 {
return offsetLine, nil, errors.New("valid operation name wasn't found in log") return offsetLine, nil, errors.New("valid operation name wasn't found in log")
} }
@ -221,7 +221,7 @@ func (lc *logConnector) getMetric(offsetLine int, name string) (int, *logMetric,
} }
// Example value = Thu Apr 13 13:28:40 2023:INFO:Device Status:: 2 VFs // Example value = Thu Apr 13 13:28:40 2023:INFO:Device Status:: 2 VFs
func (lc *logConnector) parseNumVFs(s string) (int, error) { func parseNumVFs(s string) (int, error) {
i := strings.LastIndex(s, deviceStatusStartPrefix) i := strings.LastIndex(s, deviceStatusStartPrefix)
if i == -1 { if i == -1 {
return 0, errors.New("couldn't find device status prefix in line") return 0, errors.New("couldn't find device status prefix in line")
@ -244,7 +244,7 @@ func (lc *logConnector) parseNumVFs(s string) (int, error) {
// Parse Operation name // Parse Operation name
// Example = Thu Apr 13 13:28:40 2023:INFO:5GUL counters: Code Blocks // Example = Thu Apr 13 13:28:40 2023:INFO:5GUL counters: Code Blocks
// Output: 5GUL // Output: 5GUL
func (lc *logConnector) parseOperationName(s string) string { func parseOperationName(s string) string {
i := strings.Index(s, infoLine) i := strings.Index(s, infoLine)
if i >= 0 { if i >= 0 {
j := strings.Index(s[i:], countersLine) j := strings.Index(s[i:], countersLine)

View File

@ -240,11 +240,9 @@ func TestParseOperationName(t *testing.T) {
{"", ""}, {"", ""},
} }
logConnector := prepareLogConnMock()
require.NotNil(t, logConnector)
for _, tc := range testCases { for _, tc := range testCases {
t.Run("expected "+tc.expected, func(t *testing.T) { t.Run("expected "+tc.expected, func(t *testing.T) {
operationName := logConnector.parseOperationName(tc.input) operationName := parseOperationName(tc.input)
require.Equal(t, tc.expected, operationName) require.Equal(t, tc.expected, operationName)
}) })
} }

View File

@ -50,7 +50,7 @@ type IntelDLB struct {
maxInitMessageLength uint32 maxInitMessageLength uint32
} }
func (d *IntelDLB) SampleConfig() string { func (*IntelDLB) SampleConfig() string {
return sampleConfig return sampleConfig
} }

View File

@ -56,7 +56,7 @@ type fileInfo struct {
pciBdf string // PCI Bus:Device.Function (BDF) pciBdf string // PCI Bus:Device.Function (BDF)
} }
func (p *IntelPMT) SampleConfig() string { func (*IntelPMT) SampleConfig() string {
return sampleConfig return sampleConfig
} }

View File

@ -547,9 +547,9 @@ type fakeFileInfo struct {
fileMode os.FileMode fileMode os.FileMode
} }
func (f fakeFileInfo) Name() string { return "" } func (fakeFileInfo) Name() string { return "" }
func (f fakeFileInfo) Size() int64 { return 0 } func (fakeFileInfo) Size() int64 { return 0 }
func (f fakeFileInfo) Mode() os.FileMode { return f.fileMode } func (f fakeFileInfo) Mode() os.FileMode { return f.fileMode }
func (f fakeFileInfo) ModTime() time.Time { return time.Time{} } func (fakeFileInfo) ModTime() time.Time { return time.Time{} }
func (f fakeFileInfo) IsDir() bool { return false } func (fakeFileInfo) IsDir() bool { return false }
func (f fakeFileInfo) Sys() interface{} { return nil } func (fakeFileInfo) Sys() interface{} { return nil }

View File

@ -33,7 +33,7 @@ type optGenerator struct{}
// generate takes plugin configuration options and generates options needed // generate takes plugin configuration options and generates options needed
// to gather requested metrics. // to gather requested metrics.
func (g *optGenerator) generate(cfg optConfig) []ptel.Option { func (*optGenerator) generate(cfg optConfig) []ptel.Option {
opts := make([]ptel.Option, 0) opts := make([]ptel.Option, 0)
if len(cfg.includedCPUs) != 0 { if len(cfg.includedCPUs) != 0 {
opts = append(opts, ptel.WithIncludedCPUs(cfg.includedCPUs)) opts = append(opts, ptel.WithIncludedCPUs(cfg.includedCPUs))

View File

@ -100,7 +100,7 @@ func (r *IntelRDT) Start(acc telegraf.Accumulator) error {
return nil return nil
} }
func (r *IntelRDT) Gather(_ telegraf.Accumulator) error { func (*IntelRDT) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -12,7 +12,7 @@ import (
type mockProc struct{} type mockProc struct{}
func (m *mockProc) getAllProcesses() ([]process, error) { func (*mockProc) getAllProcesses() ([]process, error) {
procs := []process{ procs := []process{
{Name: "process", PID: 1000}, {Name: "process", PID: 1000},
{Name: "process2", PID: 1002}, {Name: "process2", PID: 1002},

View File

@ -19,7 +19,7 @@ func newProcessor() processesHandler {
return &processManager{} return &processManager{}
} }
func (p *processManager) getAllProcesses() ([]process, error) { func (*processManager) getAllProcesses() ([]process, error) {
allProcesses, err := procfs.AllProcs() allProcesses, err := procfs.AllProcs()
if err != nil { if err != nil {
return nil, err return nil, err

View File

@ -179,7 +179,7 @@ func (m *Ipmi) parse(acc telegraf.Accumulator, server, sensor string) error {
return m.parseV1(acc, hostname, out, timestamp) return m.parseV1(acc, hostname, out, timestamp)
} }
case "chassis_power_status": case "chassis_power_status":
return m.parseChassisPowerStatus(acc, hostname, out, timestamp) return parseChassisPowerStatus(acc, hostname, out, timestamp)
case "dcmi_power_reading": case "dcmi_power_reading":
return m.parseDCMIPowerReading(acc, hostname, out, timestamp) return m.parseDCMIPowerReading(acc, hostname, out, timestamp)
} }
@ -187,7 +187,7 @@ func (m *Ipmi) parse(acc telegraf.Accumulator, server, sensor string) error {
return fmt.Errorf("unknown sensor type %q", sensor) return fmt.Errorf("unknown sensor type %q", sensor)
} }
func (m *Ipmi) parseChassisPowerStatus(acc telegraf.Accumulator, hostname string, cmdOut []byte, measuredAt time.Time) error { func parseChassisPowerStatus(acc telegraf.Accumulator, hostname string, cmdOut []byte, measuredAt time.Time) error {
// each line will look something like // each line will look something like
// Chassis Power is on // Chassis Power is on
// Chassis Power is off // Chassis Power is off

View File

@ -820,14 +820,10 @@ func Test_parsePowerStatus(t *testing.T) {
}, },
} }
ipmi := &Ipmi{
Log: testutil.Logger{},
}
for _, tt := range tests { for _, tt := range tests {
t.Run(tt.name, func(t *testing.T) { t.Run(tt.name, func(t *testing.T) {
var acc testutil.Accumulator var acc testutil.Accumulator
err := ipmi.parseChassisPowerStatus(&acc, tt.args.hostname, tt.args.cmdOut, tt.args.measuredAt) err := parseChassisPowerStatus(&acc, tt.args.hostname, tt.args.cmdOut, tt.args.measuredAt)
require.NoError(t, err) require.NoError(t, err)
testutil.RequireMetricsEqual(t, tt.expected, acc.GetTelegrafMetrics(), testutil.IgnoreTime()) testutil.RequireMetricsEqual(t, tt.expected, acc.GetTelegrafMetrics(), testutil.IgnoreTime())
}) })

View File

@ -38,7 +38,7 @@ func (*Ipset) SampleConfig() string {
return sampleConfig return sampleConfig
} }
func (i *Ipset) Init() error { func (*Ipset) Init() error {
_, err := exec.LookPath("ipset") _, err := exec.LookPath("ipset")
if err != nil { if err != nil {
return err return err

View File

@ -172,7 +172,7 @@ func (m *OpenConfigTelemetry) Start(acc telegraf.Accumulator) error {
return nil return nil
} }
func (m *OpenConfigTelemetry) Gather(_ telegraf.Accumulator) error { func (*OpenConfigTelemetry) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -58,10 +58,7 @@ type openConfigTelemetryServer struct {
telemetry.UnimplementedOpenConfigTelemetryServer telemetry.UnimplementedOpenConfigTelemetryServer
} }
func (s *openConfigTelemetryServer) TelemetrySubscribe( func (*openConfigTelemetryServer) TelemetrySubscribe(req *telemetry.SubscriptionRequest, stream telemetry.OpenConfigTelemetry_TelemetrySubscribeServer) error {
req *telemetry.SubscriptionRequest,
stream telemetry.OpenConfigTelemetry_TelemetrySubscribeServer,
) error {
path := req.PathList[0].Path path := req.PathList[0].Path
switch path { switch path {
case "/sensor": case "/sensor":
@ -78,28 +75,28 @@ func (s *openConfigTelemetryServer) TelemetrySubscribe(
return nil return nil
} }
func (s *openConfigTelemetryServer) CancelTelemetrySubscription( func (*openConfigTelemetryServer) CancelTelemetrySubscription(
_ context.Context, context.Context,
_ *telemetry.CancelSubscriptionRequest, *telemetry.CancelSubscriptionRequest,
) (*telemetry.CancelSubscriptionReply, error) { ) (*telemetry.CancelSubscriptionReply, error) {
return nil, nil return nil, nil
} }
func (s *openConfigTelemetryServer) GetTelemetrySubscriptions( func (*openConfigTelemetryServer) GetTelemetrySubscriptions(
_ context.Context, context.Context,
_ *telemetry.GetSubscriptionsRequest, *telemetry.GetSubscriptionsRequest,
) (*telemetry.GetSubscriptionsReply, error) { ) (*telemetry.GetSubscriptionsReply, error) {
return nil, nil return nil, nil
} }
func (s *openConfigTelemetryServer) GetTelemetryOperationalState( func (*openConfigTelemetryServer) GetTelemetryOperationalState(
_ context.Context, context.Context,
_ *telemetry.GetOperationalStateRequest, *telemetry.GetOperationalStateRequest,
) (*telemetry.GetOperationalStateReply, error) { ) (*telemetry.GetOperationalStateReply, error) {
return nil, nil return nil, nil
} }
func (s *openConfigTelemetryServer) GetDataEncodings(_ context.Context, _ *telemetry.DataEncodingRequest) (*telemetry.DataEncodingReply, error) { func (*openConfigTelemetryServer) GetDataEncodings(context.Context, *telemetry.DataEncodingRequest) (*telemetry.DataEncodingReply, error) {
return nil, nil return nil, nil
} }

View File

@ -299,7 +299,7 @@ func (k *KafkaConsumer) Start(acc telegraf.Accumulator) error {
return nil return nil
} }
func (k *KafkaConsumer) Gather(_ telegraf.Accumulator) error { func (*KafkaConsumer) Gather(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -240,53 +240,53 @@ type FakeConsumerGroupSession struct {
ctx context.Context ctx context.Context
} }
func (s *FakeConsumerGroupSession) Claims() map[string][]int32 { func (*FakeConsumerGroupSession) Claims() map[string][]int32 {
panic("not implemented") panic("not implemented")
} }
func (s *FakeConsumerGroupSession) MemberID() string { func (*FakeConsumerGroupSession) MemberID() string {
panic("not implemented") panic("not implemented")
} }
func (s *FakeConsumerGroupSession) GenerationID() int32 { func (*FakeConsumerGroupSession) GenerationID() int32 {
panic("not implemented") panic("not implemented")
} }
func (s *FakeConsumerGroupSession) MarkOffset(_ string, _ int32, _ int64, _ string) { func (*FakeConsumerGroupSession) MarkOffset(string, int32, int64, string) {
panic("not implemented") panic("not implemented")
} }
func (s *FakeConsumerGroupSession) ResetOffset(_ string, _ int32, _ int64, _ string) { func (*FakeConsumerGroupSession) ResetOffset(string, int32, int64, string) {
panic("not implemented") panic("not implemented")
} }
func (s *FakeConsumerGroupSession) MarkMessage(_ *sarama.ConsumerMessage, _ string) { func (*FakeConsumerGroupSession) MarkMessage(*sarama.ConsumerMessage, string) {
} }
func (s *FakeConsumerGroupSession) Context() context.Context { func (s *FakeConsumerGroupSession) Context() context.Context {
return s.ctx return s.ctx
} }
func (s *FakeConsumerGroupSession) Commit() { func (*FakeConsumerGroupSession) Commit() {
} }
type FakeConsumerGroupClaim struct { type FakeConsumerGroupClaim struct {
messages chan *sarama.ConsumerMessage messages chan *sarama.ConsumerMessage
} }
func (c *FakeConsumerGroupClaim) Topic() string { func (*FakeConsumerGroupClaim) Topic() string {
panic("not implemented") panic("not implemented")
} }
func (c *FakeConsumerGroupClaim) Partition() int32 { func (*FakeConsumerGroupClaim) Partition() int32 {
panic("not implemented") panic("not implemented")
} }
func (c *FakeConsumerGroupClaim) InitialOffset() int64 { func (*FakeConsumerGroupClaim) InitialOffset() int64 {
panic("not implemented") panic("not implemented")
} }
func (c *FakeConsumerGroupClaim) HighWaterMarkOffset() int64 { func (*FakeConsumerGroupClaim) HighWaterMarkOffset() int64 {
panic("not implemented") panic("not implemented")
} }

View File

@ -68,12 +68,12 @@ func (k *Kernel) Init() error {
} }
func (k *Kernel) Gather(acc telegraf.Accumulator) error { func (k *Kernel) Gather(acc telegraf.Accumulator) error {
data, err := k.getProcValueBytes(k.statFile) data, err := getProcValueBytes(k.statFile)
if err != nil { if err != nil {
return err return err
} }
entropyValue, err := k.getProcValueInt(k.entropyStatFile) entropyValue, err := getProcValueInt(k.entropyStatFile)
if err != nil { if err != nil {
return err return err
} }
@ -137,7 +137,7 @@ func (k *Kernel) Gather(acc telegraf.Accumulator) error {
extraStats := []string{"general_profit"} extraStats := []string{"general_profit"}
for _, f := range stats { for _, f := range stats {
m, err := k.getProcValueInt(filepath.Join(k.ksmStatsDir, f)) m, err := getProcValueInt(filepath.Join(k.ksmStatsDir, f))
if err != nil { if err != nil {
return err return err
} }
@ -146,7 +146,7 @@ func (k *Kernel) Gather(acc telegraf.Accumulator) error {
} }
for _, f := range extraStats { for _, f := range extraStats {
m, err := k.getProcValueInt(filepath.Join(k.ksmStatsDir, f)) m, err := getProcValueInt(filepath.Join(k.ksmStatsDir, f))
if err != nil { if err != nil {
// if an extraStats metric doesn't exist in our kernel version, ignore it. // if an extraStats metric doesn't exist in our kernel version, ignore it.
continue continue
@ -166,7 +166,7 @@ func (k *Kernel) Gather(acc telegraf.Accumulator) error {
return nil return nil
} }
func (k *Kernel) getProcValueBytes(path string) ([]byte, error) { func getProcValueBytes(path string) ([]byte, error) {
if _, err := os.Stat(path); os.IsNotExist(err) { if _, err := os.Stat(path); os.IsNotExist(err) {
return nil, fmt.Errorf("path %q does not exist", path) return nil, fmt.Errorf("path %q does not exist", path)
} else if err != nil { } else if err != nil {
@ -181,8 +181,8 @@ func (k *Kernel) getProcValueBytes(path string) ([]byte, error) {
return data, nil return data, nil
} }
func (k *Kernel) getProcValueInt(path string) (int64, error) { func getProcValueInt(path string) (int64, error) {
data, err := k.getProcValueBytes(path) data, err := getProcValueBytes(path)
if err != nil { if err != nil {
return -1, err return -1, err
} }

View File

@ -14,23 +14,13 @@ import (
) )
func TestGetProcValueInt(t *testing.T) { func TestGetProcValueInt(t *testing.T) {
k := Kernel{ d, err := getProcValueInt("testdata/entropy_stat_file_full")
statFile: "testdata/stat_file_full",
entropyStatFile: "testdata/entropy_stat_file_full",
}
d, err := k.getProcValueInt(k.entropyStatFile)
require.NoError(t, err) require.NoError(t, err)
require.IsType(t, int64(1), d) require.IsType(t, int64(1), d)
} }
func TestGetProcValueByte(t *testing.T) { func TestGetProcValueByte(t *testing.T) {
k := Kernel{ d, err := getProcValueBytes("testdata/entropy_stat_file_full")
statFile: "testdata/stat_file_full",
entropyStatFile: "testdata/entropy_stat_file_full",
}
d, err := k.getProcValueBytes(k.entropyStatFile)
require.NoError(t, err) require.NoError(t, err)
require.IsType(t, []byte("test"), d) require.IsType(t, []byte("test"), d)
} }

View File

@ -101,7 +101,7 @@ func (*Kibana) SampleConfig() string {
return sampleConfig return sampleConfig
} }
func (k *Kibana) Start(_ telegraf.Accumulator) error { func (*Kibana) Start(telegraf.Accumulator) error {
return nil return nil
} }

View File

@ -19,7 +19,7 @@ func collectSecrets(ctx context.Context, acc telegraf.Accumulator, ki *Kubernete
return return
} }
for _, i := range list.Items { for _, i := range list.Items {
ki.gatherCertificates(i, acc) gatherCertificates(i, acc)
} }
} }
@ -59,7 +59,7 @@ func getTags(cert *x509.Certificate) map[string]string {
return tags return tags
} }
func (ki *KubernetesInventory) gatherCertificates(r corev1.Secret, acc telegraf.Accumulator) { func gatherCertificates(r corev1.Secret, acc telegraf.Accumulator) {
now := time.Now() now := time.Now()
for resourceName, val := range r.Data { for resourceName, val := range r.Data {

View File

@ -15,11 +15,11 @@ func collectEndpoints(ctx context.Context, acc telegraf.Accumulator, ki *Kuberne
return return
} }
for _, i := range list.Items { for _, i := range list.Items {
ki.gatherEndpoint(i, acc) gatherEndpoint(i, acc)
} }
} }
func (ki *KubernetesInventory) gatherEndpoint(e corev1.Endpoints, acc telegraf.Accumulator) { func gatherEndpoint(e corev1.Endpoints, acc telegraf.Accumulator) {
creationTs := e.GetCreationTimestamp() creationTs := e.GetCreationTimestamp()
if creationTs.IsZero() { if creationTs.IsZero() {
return return

View File

@ -13,8 +13,6 @@ import (
) )
func TestEndpoint(t *testing.T) { func TestEndpoint(t *testing.T) {
cli := &client{}
now := time.Now() now := time.Now()
now = time.Date(now.Year(), now.Month(), now.Day(), now.Hour(), 1, 36, 0, now.Location()) now = time.Date(now.Year(), now.Month(), now.Day(), now.Hour(), 1, 36, 0, now.Location())
@ -256,12 +254,9 @@ func TestEndpoint(t *testing.T) {
} }
for _, v := range tests { for _, v := range tests {
ks := &KubernetesInventory{
client: cli,
}
acc := new(testutil.Accumulator) acc := new(testutil.Accumulator)
for _, endpoint := range ((v.handler.responseMap["/endpoints/"]).(*v1.EndpointsList)).Items { for _, endpoint := range ((v.handler.responseMap["/endpoints/"]).(*v1.EndpointsList)).Items {
ks.gatherEndpoint(endpoint, acc) gatherEndpoint(endpoint, acc)
} }
err := acc.FirstError() err := acc.FirstError()

View File

@ -15,11 +15,11 @@ func collectIngress(ctx context.Context, acc telegraf.Accumulator, ki *Kubernete
return return
} }
for _, i := range list.Items { for _, i := range list.Items {
ki.gatherIngress(i, acc) gatherIngress(i, acc)
} }
} }
func (ki *KubernetesInventory) gatherIngress(i netv1.Ingress, acc telegraf.Accumulator) { func gatherIngress(i netv1.Ingress, acc telegraf.Accumulator) {
creationTs := i.GetCreationTimestamp() creationTs := i.GetCreationTimestamp()
if creationTs.IsZero() { if creationTs.IsZero() {
return return

View File

@ -13,8 +13,6 @@ import (
) )
func TestIngress(t *testing.T) { func TestIngress(t *testing.T) {
cli := &client{}
now := time.Now() now := time.Now()
now = time.Date(now.Year(), now.Month(), now.Day(), now.Hour(), 1, 36, 0, now.Location()) now = time.Date(now.Year(), now.Month(), now.Day(), now.Hour(), 1, 36, 0, now.Location())
@ -219,12 +217,9 @@ func TestIngress(t *testing.T) {
} }
for _, v := range tests { for _, v := range tests {
ks := &KubernetesInventory{
client: cli,
}
acc := new(testutil.Accumulator) acc := new(testutil.Accumulator)
for _, ingress := range ((v.handler.responseMap["/ingress/"]).(netv1.IngressList)).Items { for _, ingress := range ((v.handler.responseMap["/ingress/"]).(netv1.IngressList)).Items {
ks.gatherIngress(ingress, acc) gatherIngress(ingress, acc)
} }
err := acc.FirstError() err := acc.FirstError()

View File

@ -15,14 +15,14 @@ func collectNodes(ctx context.Context, acc telegraf.Accumulator, ki *KubernetesI
return return
} }
ki.gatherNodeCount(len(list.Items), acc) gatherNodeCount(len(list.Items), acc)
for i := range list.Items { for i := range list.Items {
ki.gatherNode(&list.Items[i], acc) ki.gatherNode(&list.Items[i], acc)
} }
} }
func (ki *KubernetesInventory) gatherNodeCount(count int, acc telegraf.Accumulator) { func gatherNodeCount(count int, acc telegraf.Accumulator) {
fields := map[string]interface{}{"node_count": count} fields := map[string]interface{}{"node_count": count}
tags := make(map[string]string) tags := make(map[string]string)

View File

@ -173,7 +173,7 @@ func TestNode(t *testing.T) {
if v.name == "no nodes" { if v.name == "no nodes" {
nodeCount := len((v.handler.responseMap["/nodes/"]).(corev1.NodeList).Items) nodeCount := len((v.handler.responseMap["/nodes/"]).(corev1.NodeList).Items)
ks.gatherNodeCount(nodeCount, acc) gatherNodeCount(nodeCount, acc)
} }
require.Len(t, acc.Metrics, len(v.output)) require.Len(t, acc.Metrics, len(v.output))
testutil.RequireMetricsEqual(t, acc.GetTelegrafMetrics(), v.output, testutil.IgnoreTime()) testutil.RequireMetricsEqual(t, acc.GetTelegrafMetrics(), v.output, testutil.IgnoreTime())

View File

@ -16,11 +16,11 @@ func collectPersistentVolumes(ctx context.Context, acc telegraf.Accumulator, ki
return return
} }
for i := range list.Items { for i := range list.Items {
ki.gatherPersistentVolume(&list.Items[i], acc) gatherPersistentVolume(&list.Items[i], acc)
} }
} }
func (ki *KubernetesInventory) gatherPersistentVolume(pv *corev1.PersistentVolume, acc telegraf.Accumulator) { func gatherPersistentVolume(pv *corev1.PersistentVolume, acc telegraf.Accumulator) {
phaseType := 5 phaseType := 5
switch strings.ToLower(string(pv.Status.Phase)) { switch strings.ToLower(string(pv.Status.Phase)) {
case "bound": case "bound":

View File

@ -13,7 +13,6 @@ import (
) )
func TestPersistentVolume(t *testing.T) { func TestPersistentVolume(t *testing.T) {
cli := &client{}
now := time.Now() now := time.Now()
now = time.Date(now.Year(), now.Month(), now.Day(), now.Hour(), 1, 36, 0, now.Location()) now = time.Date(now.Year(), now.Month(), now.Day(), now.Hour(), 1, 36, 0, now.Location())
@ -77,13 +76,10 @@ func TestPersistentVolume(t *testing.T) {
} }
for _, v := range tests { for _, v := range tests {
ks := &KubernetesInventory{
client: cli,
}
acc := new(testutil.Accumulator) acc := new(testutil.Accumulator)
items := ((v.handler.responseMap["/persistentvolumes/"]).(*corev1.PersistentVolumeList)).Items items := ((v.handler.responseMap["/persistentvolumes/"]).(*corev1.PersistentVolumeList)).Items
for i := range items { for i := range items {
ks.gatherPersistentVolume(&items[i], acc) gatherPersistentVolume(&items[i], acc)
} }
err := acc.FirstError() err := acc.FirstError()