Update changelog
(cherry picked from commit e9872741674c054b58a4236d3c3ea98f18515c65)
This commit is contained in:
parent
7de9c5ff27
commit
f988d77049
53
CHANGELOG.md
53
CHANGELOG.md
|
|
@ -1,3 +1,56 @@
|
|||
## v1.20.0-rc0 [2021-09-02]
|
||||
|
||||
#### Release Notes
|
||||
|
||||
- [#9642](https://github.com/influxdata/telegraf/pull/9642) Build with Golang 1.17
|
||||
|
||||
#### Bugfixes
|
||||
|
||||
- [#9700](https://github.com/influxdata/telegraf/pull/9700) Update thrift module to 0.14.2 and zipkin-go-opentracing 0.4.5
|
||||
- [#9587](https://github.com/influxdata/telegraf/pull/9587) `outputs.opentelemetry` Use headers config in grpc requests
|
||||
- [#9713](https://github.com/influxdata/telegraf/pull/9713) Update runc module to v1.0.0-rc95 to address CVE-2021-30465
|
||||
- [#9699](https://github.com/influxdata/telegraf/pull/9699) Migrate dgrijalva/jwt-go to golang-jwt/jwt/v4
|
||||
- [#9139](https://github.com/influxdata/telegraf/pull/9139) `serializers.prometheus` Update timestamps and expiration time as new data arrives
|
||||
- [#9625](https://github.com/influxdata/telegraf/pull/9625) `outputs.graylog` Output timestamp with fractional seconds
|
||||
- [#9655](https://github.com/influxdata/telegraf/pull/9655) Update cloud.google.com/go/pubsub module from 1.2.0 to 1.15.0
|
||||
- [#9674](https://github.com/influxdata/telegraf/pull/9674) `inputs.mongodb` Change command based on server version
|
||||
- [#9676](https://github.com/influxdata/telegraf/pull/9676) `outputs.dynatrace` Remove hardcoded int value
|
||||
- [#9619](https://github.com/influxdata/telegraf/pull/9619) `outputs.influxdb_v2` Increase accepted retry-after header values.
|
||||
- [#9652](https://github.com/influxdata/telegraf/pull/9652) Update github.com/tinylib/msgp module from 1.1.5 to 1.1.6
|
||||
- [#9471](https://github.com/influxdata/telegraf/pull/9471) `inputs.sql` Make timeout apply to single query
|
||||
|
||||
#### Features
|
||||
|
||||
- [#9665](https://github.com/influxdata/telegraf/pull/9665) `inputs.systemd_units` feat(plugins/inputs/systemd_units): add pattern support
|
||||
- [#9598](https://github.com/influxdata/telegraf/pull/9598) `outputs.sql` Add bool datatype
|
||||
- [#9386](https://github.com/influxdata/telegraf/pull/9386) `inputs.cloudwatch` Pull metrics from multiple AWS CloudWatch namespaces
|
||||
- [#9411](https://github.com/influxdata/telegraf/pull/9411) `inputs.cloudwatch` Support AWS Web Identity Provider
|
||||
- [#9570](https://github.com/influxdata/telegraf/pull/9570) `inputs.modbus` Add support for RTU over TCP
|
||||
- [#9488](https://github.com/influxdata/telegraf/pull/9488) `inputs.procstat` Support cgroup globs and include systemd unit children
|
||||
- [#9322](https://github.com/influxdata/telegraf/pull/9322) `inputs.suricata` Support alert event type
|
||||
- [#5464](https://github.com/influxdata/telegraf/pull/5464) `inputs.prometheus` Add ability to query Consul Service catalog
|
||||
- [#8641](https://github.com/influxdata/telegraf/pull/8641) `outputs.prometheus_client` Add Landing page
|
||||
- [#9529](https://github.com/influxdata/telegraf/pull/9529) `inputs.http_listener_v2` Allows multiple paths and add path_tag
|
||||
- [#9395](https://github.com/influxdata/telegraf/pull/9395) Add cookie authentication to HTTP input and output plugins
|
||||
- [#8454](https://github.com/influxdata/telegraf/pull/8454) `inputs.syslog` Add RFC3164 support
|
||||
- [#9351](https://github.com/influxdata/telegraf/pull/9351) `inputs.jenkins` Add option to include nodes by name
|
||||
- [#9277](https://github.com/influxdata/telegraf/pull/9277) Add JSON, MessagePack, and Protocol-buffers format support to the XPath parser
|
||||
- [#9343](https://github.com/influxdata/telegraf/pull/9343) `inputs.snmp_trap` Improve MIB lookup performance
|
||||
- [#9342](https://github.com/influxdata/telegraf/pull/9342) `outputs.newrelic` Add option to override metric_url
|
||||
- [#9306](https://github.com/influxdata/telegraf/pull/9306) `inputs.smart` Add power mode status
|
||||
|
||||
#### New Input Plugins
|
||||
|
||||
- [#9602](https://github.com/influxdata/telegraf/pull/9602) Add rocm_smi input to monitor AMD GPUs
|
||||
- [#9101](https://github.com/influxdata/telegraf/pull/9101) Add mdstat input to gather from /proc/mdstat collection
|
||||
- [#3536](https://github.com/influxdata/telegraf/pull/3536) Add Elasticsearch query input
|
||||
|
||||
#### New Output Plugins
|
||||
|
||||
- [#9228](https://github.com/influxdata/telegraf/pull/9228) Add OpenTelemetry output
|
||||
- [#9426](https://github.com/influxdata/telegraf/pull/9426) Add Azure Data Explorer(ADX) output
|
||||
|
||||
|
||||
## v1.19.3 [2021-08-18]
|
||||
|
||||
#### Bugfixes
|
||||
|
|
|
|||
|
|
@ -315,6 +315,30 @@
|
|||
# # "ai.cloud.roleInstance" = "kubernetes_pod_name"
|
||||
|
||||
|
||||
# # Sends metrics to Azure Data Explorer
|
||||
# [[outputs.azure_data_explorer]]
|
||||
# ## Azure Data Exlorer cluster endpoint
|
||||
# ## ex: endpoint_url = "https://clustername.australiasoutheast.kusto.windows.net"
|
||||
# endpoint_url = ""
|
||||
#
|
||||
# ## The Azure Data Explorer database that the metrics will be ingested into.
|
||||
# ## The plugin will NOT generate this database automatically, it's expected that this database already exists before ingestion.
|
||||
# ## ex: "exampledatabase"
|
||||
# database = ""
|
||||
#
|
||||
# ## Timeout for Azure Data Explorer operations
|
||||
# # timeout = "20s"
|
||||
#
|
||||
# ## Type of metrics grouping used when pushing to Azure Data Explorer.
|
||||
# ## Default is "TablePerMetric" for one table per different metric.
|
||||
# ## For more information, please check the plugin README.
|
||||
# # metrics_grouping_type = "TablePerMetric"
|
||||
#
|
||||
# ## Name of the single table to store all the metrics (Only needed if metrics_grouping_type is "SingleTable").
|
||||
# # table_name = ""
|
||||
#
|
||||
|
||||
|
||||
# # Send aggregate metrics to Azure Monitor
|
||||
# [[outputs.azure_monitor]]
|
||||
# ## Timeout for HTTP writes.
|
||||
|
|
@ -404,16 +428,19 @@
|
|||
#
|
||||
# ## Amazon Credentials
|
||||
# ## Credentials are loaded in the following order
|
||||
# ## 1) Assumed credentials via STS if role_arn is specified
|
||||
# ## 2) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 3) shared profile from 'profile'
|
||||
# ## 4) environment variables
|
||||
# ## 5) shared credentials file
|
||||
# ## 6) EC2 Instance Profile
|
||||
# ## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||
# ## 2) Assumed credentials via STS if role_arn is specified
|
||||
# ## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 4) shared profile from 'profile'
|
||||
# ## 5) environment variables
|
||||
# ## 6) shared credentials file
|
||||
# ## 7) EC2 Instance Profile
|
||||
# #access_key = ""
|
||||
# #secret_key = ""
|
||||
# #token = ""
|
||||
# #role_arn = ""
|
||||
# #web_identity_token_file = ""
|
||||
# #role_session_name = ""
|
||||
# #profile = ""
|
||||
# #shared_credential_file = ""
|
||||
#
|
||||
|
|
@ -452,16 +479,19 @@
|
|||
#
|
||||
# ## Amazon Credentials
|
||||
# ## Credentials are loaded in the following order
|
||||
# ## 1) Assumed credentials via STS if role_arn is specified
|
||||
# ## 2) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 3) shared profile from 'profile'
|
||||
# ## 4) environment variables
|
||||
# ## 5) shared credentials file
|
||||
# ## 6) EC2 Instance Profile
|
||||
# ## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||
# ## 2) Assumed credentials via STS if role_arn is specified
|
||||
# ## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 4) shared profile from 'profile'
|
||||
# ## 5) environment variables
|
||||
# ## 6) shared credentials file
|
||||
# ## 7) EC2 Instance Profile
|
||||
# #access_key = ""
|
||||
# #secret_key = ""
|
||||
# #token = ""
|
||||
# #role_arn = ""
|
||||
# #web_identity_token_file = ""
|
||||
# #role_session_name = ""
|
||||
# #profile = ""
|
||||
# #shared_credential_file = ""
|
||||
#
|
||||
|
|
@ -809,6 +839,15 @@
|
|||
# ## Use TLS but skip chain & host verification
|
||||
# # insecure_skip_verify = false
|
||||
#
|
||||
# ## Optional Cookie authentication
|
||||
# # cookie_auth_url = "https://localhost/authMe"
|
||||
# # cookie_auth_method = "POST"
|
||||
# # cookie_auth_username = "username"
|
||||
# # cookie_auth_password = "pa$$word"
|
||||
# # cookie_auth_body = '{"username": "user", "password": "pa$$word", "authenticate": "me"}'
|
||||
# ## cookie_auth_renewal not set or set to "0" will auth once and never renew the cookie
|
||||
# # cookie_auth_renewal = "5m"
|
||||
#
|
||||
# ## Data format to output.
|
||||
# ## Each data format has it's own unique set of configuration options, read
|
||||
# ## more about them here:
|
||||
|
|
@ -1049,16 +1088,19 @@
|
|||
#
|
||||
# ## Amazon Credentials
|
||||
# ## Credentials are loaded in the following order
|
||||
# ## 1) Assumed credentials via STS if role_arn is specified
|
||||
# ## 2) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 3) shared profile from 'profile'
|
||||
# ## 4) environment variables
|
||||
# ## 5) shared credentials file
|
||||
# ## 6) EC2 Instance Profile
|
||||
# ## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||
# ## 2) Assumed credentials via STS if role_arn is specified
|
||||
# ## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 4) shared profile from 'profile'
|
||||
# ## 5) environment variables
|
||||
# ## 6) shared credentials file
|
||||
# ## 7) EC2 Instance Profile
|
||||
# #access_key = ""
|
||||
# #secret_key = ""
|
||||
# #token = ""
|
||||
# #role_arn = ""
|
||||
# #web_identity_token_file = ""
|
||||
# #role_session_name = ""
|
||||
# #profile = ""
|
||||
# #shared_credential_file = ""
|
||||
#
|
||||
|
|
@ -1268,6 +1310,10 @@
|
|||
# ## HTTP Proxy override. If unset use values from the standard
|
||||
# ## proxy environment variables to determine proxy, if any.
|
||||
# # http_proxy = "http://corporate.proxy:3128"
|
||||
#
|
||||
# ## Metric URL override to enable geographic location endpoints.
|
||||
# # If not set use values from the standard
|
||||
# # metric_url = "https://metric-api.newrelic.com/metric/v1"
|
||||
|
||||
|
||||
# # Send telegraf measurements to NSQD
|
||||
|
|
@ -1284,6 +1330,41 @@
|
|||
# data_format = "influx"
|
||||
|
||||
|
||||
# # Send OpenTelemetry metrics over gRPC
|
||||
# [[outputs.opentelemetry]]
|
||||
# ## Override the default (localhost:4317) OpenTelemetry gRPC service
|
||||
# ## address:port
|
||||
# # service_address = "localhost:4317"
|
||||
#
|
||||
# ## Override the default (5s) request timeout
|
||||
# # timeout = "5s"
|
||||
#
|
||||
# ## Optional TLS Config.
|
||||
# ##
|
||||
# ## Root certificates for verifying server certificates encoded in PEM format.
|
||||
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||
# ## The public and private keypairs for the client encoded in PEM format.
|
||||
# ## May contain intermediate certificates.
|
||||
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||
# # tls_key = "/etc/telegraf/key.pem"
|
||||
# ## Use TLS, but skip TLS chain and host verification.
|
||||
# # insecure_skip_verify = false
|
||||
# ## Send the specified TLS server name via SNI.
|
||||
# # tls_server_name = "foo.example.com"
|
||||
#
|
||||
# ## Override the default (gzip) compression used to send data.
|
||||
# ## Supports: "gzip", "none"
|
||||
# # compression = "gzip"
|
||||
#
|
||||
# ## Additional OpenTelemetry resource attributes
|
||||
# # [outputs.opentelemetry.attributes]
|
||||
# # "service.name" = "demo"
|
||||
#
|
||||
# ## Additional gRPC request metadata
|
||||
# # [outputs.opentelemetry.headers]
|
||||
# # key1 = "value1"
|
||||
|
||||
|
||||
# # Configuration for OpenTSDB server to send metrics to
|
||||
# [[outputs.opentsdb]]
|
||||
# ## prefix for metrics keys
|
||||
|
|
@ -1748,16 +1829,19 @@
|
|||
#
|
||||
# ## Amazon Credentials
|
||||
# ## Credentials are loaded in the following order:
|
||||
# ## 1) Assumed credentials via STS if role_arn is specified
|
||||
# ## 2) Explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 3) Shared profile from 'profile'
|
||||
# ## 4) Environment variables
|
||||
# ## 5) Shared credentials file
|
||||
# ## 6) EC2 Instance Profile
|
||||
# ## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||
# ## 2) Assumed credentials via STS if role_arn is specified
|
||||
# ## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 4) shared profile from 'profile'
|
||||
# ## 5) environment variables
|
||||
# ## 6) shared credentials file
|
||||
# ## 7) EC2 Instance Profile
|
||||
# #access_key = ""
|
||||
# #secret_key = ""
|
||||
# #token = ""
|
||||
# #role_arn = ""
|
||||
# #web_identity_token_file = ""
|
||||
# #role_session_name = ""
|
||||
# #profile = ""
|
||||
# #shared_credential_file = ""
|
||||
#
|
||||
|
|
@ -1870,7 +1954,7 @@
|
|||
# ## Print Warp 10 error body
|
||||
# # print_error_body = false
|
||||
#
|
||||
# ## Max string error size
|
||||
# ## Max string error size
|
||||
# # max_string_error_size = 511
|
||||
#
|
||||
# ## Optional TLS Config
|
||||
|
|
@ -2896,6 +2980,15 @@
|
|||
# # num_histogram_buckets = 100 # default: 10
|
||||
|
||||
|
||||
# # Query statistics from AMD Graphics cards using rocm-smi binary
|
||||
# [[inputs.amd_rocm_smi]]
|
||||
# ## Optional: path to rocm-smi binary, defaults to $PATH via exec.LookPath
|
||||
# # bin_path = "/opt/rocm/bin/rocm-smi"
|
||||
#
|
||||
# ## Optional: timeout for GPU polling
|
||||
# # timeout = "5s"
|
||||
|
||||
|
||||
# # Read Apache status information (mod_status)
|
||||
# [[inputs.apache]]
|
||||
# ## An array of URLs to gather from, must be directed at the machine
|
||||
|
|
@ -3163,16 +3256,19 @@
|
|||
#
|
||||
# ## Amazon Credentials
|
||||
# ## Credentials are loaded in the following order
|
||||
# ## 1) Assumed credentials via STS if role_arn is specified
|
||||
# ## 2) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 3) shared profile from 'profile'
|
||||
# ## 4) environment variables
|
||||
# ## 5) shared credentials file
|
||||
# ## 6) EC2 Instance Profile
|
||||
# ## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||
# ## 2) Assumed credentials via STS if role_arn is specified
|
||||
# ## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 4) shared profile from 'profile'
|
||||
# ## 5) environment variables
|
||||
# ## 6) shared credentials file
|
||||
# ## 7) EC2 Instance Profile
|
||||
# # access_key = ""
|
||||
# # secret_key = ""
|
||||
# # token = ""
|
||||
# # role_arn = ""
|
||||
# # web_identity_token_file = ""
|
||||
# # role_session_name = ""
|
||||
# # profile = ""
|
||||
# # shared_credential_file = ""
|
||||
#
|
||||
|
|
@ -3212,8 +3308,10 @@
|
|||
# ## Configure the TTL for the internal cache of metrics.
|
||||
# # cache_ttl = "1h"
|
||||
#
|
||||
# ## Metric Statistic Namespace (required)
|
||||
# namespace = "AWS/ELB"
|
||||
# ## Metric Statistic Namespaces (required)
|
||||
# namespaces = ["AWS/ELB"]
|
||||
# # A single metric statistic namespace that will be appended to namespaces on startup
|
||||
# # namespace = "AWS/ELB"
|
||||
#
|
||||
# ## Maximum requests per second. Note that the global default AWS rate limit is
|
||||
# ## 50 reqs/sec, so if you define multiple namespaces, these should add up to a
|
||||
|
|
@ -3632,6 +3730,72 @@
|
|||
# # num_most_recent_indices = 0
|
||||
|
||||
|
||||
# # Derive metrics from aggregating Elasticsearch query results
|
||||
# [[inputs.elasticsearch_query]]
|
||||
# ## The full HTTP endpoint URL for your Elasticsearch instance
|
||||
# ## Multiple urls can be specified as part of the same cluster,
|
||||
# ## this means that only ONE of the urls will be written to each interval.
|
||||
# urls = [ "http://node1.es.example.com:9200" ] # required.
|
||||
#
|
||||
# ## Elasticsearch client timeout, defaults to "5s".
|
||||
# # timeout = "5s"
|
||||
#
|
||||
# ## Set to true to ask Elasticsearch a list of all cluster nodes,
|
||||
# ## thus it is not necessary to list all nodes in the urls config option
|
||||
# # enable_sniffer = false
|
||||
#
|
||||
# ## Set the interval to check if the Elasticsearch nodes are available
|
||||
# ## This option is only used if enable_sniffer is also set (0s to disable it)
|
||||
# # health_check_interval = "10s"
|
||||
#
|
||||
# ## HTTP basic authentication details (eg. when using x-pack)
|
||||
# # username = "telegraf"
|
||||
# # password = "mypassword"
|
||||
#
|
||||
# ## Optional TLS Config
|
||||
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||
# # tls_key = "/etc/telegraf/key.pem"
|
||||
# ## Use TLS but skip chain & host verification
|
||||
# # insecure_skip_verify = false
|
||||
#
|
||||
# [[inputs.elasticsearch_query.aggregation]]
|
||||
# ## measurement name for the results of the aggregation query
|
||||
# measurement_name = "measurement"
|
||||
#
|
||||
# ## Elasticsearch indexes to query (accept wildcards).
|
||||
# index = "index-*"
|
||||
#
|
||||
# ## The date/time field in the Elasticsearch index (mandatory).
|
||||
# date_field = "@timestamp"
|
||||
#
|
||||
# ## Time window to query (eg. "1m" to query documents from last minute).
|
||||
# ## Normally should be set to same as collection interval
|
||||
# query_period = "1m"
|
||||
#
|
||||
# ## Lucene query to filter results
|
||||
# # filter_query = "*"
|
||||
#
|
||||
# ## Fields to aggregate values (must be numeric fields)
|
||||
# # metric_fields = ["metric"]
|
||||
#
|
||||
# ## Aggregation function to use on the metric fields
|
||||
# ## Must be set if 'metric_fields' is set
|
||||
# ## Valid values are: avg, sum, min, max, sum
|
||||
# # metric_function = "avg"
|
||||
#
|
||||
# ## Fields to be used as tags
|
||||
# ## Must be text, non-analyzed fields. Metric aggregations are performed per tag
|
||||
# # tags = ["field.keyword", "field2.keyword"]
|
||||
#
|
||||
# ## Set to true to not ignore documents when the tag(s) above are missing
|
||||
# # include_missing_tag = false
|
||||
#
|
||||
# ## String value of the tag when the tag does not exist
|
||||
# ## Used when include_missing_tag is true
|
||||
# # missing_tag_value = "null"
|
||||
|
||||
|
||||
# # Returns ethtool statistics for given interfaces
|
||||
# [[inputs.ethtool]]
|
||||
# ## List of interfaces to pull metrics for
|
||||
|
|
@ -3944,6 +4108,15 @@
|
|||
# ## Use TLS but skip chain & host verification
|
||||
# # insecure_skip_verify = false
|
||||
#
|
||||
# ## Optional Cookie authentication
|
||||
# # cookie_auth_url = "https://localhost/authMe"
|
||||
# # cookie_auth_method = "POST"
|
||||
# # cookie_auth_username = "username"
|
||||
# # cookie_auth_password = "pa$$word"
|
||||
# # cookie_auth_body = '{"username": "user", "password": "pa$$word", "authenticate": "me"}'
|
||||
# ## cookie_auth_renewal not set or set to "0" will auth once and never renew the cookie
|
||||
# # cookie_auth_renewal = "5m"
|
||||
#
|
||||
# ## Amount of time allowed to complete the HTTP request
|
||||
# # timeout = "5s"
|
||||
#
|
||||
|
|
@ -4286,7 +4459,9 @@
|
|||
# # job_include = [ "*" ]
|
||||
# # job_exclude = [ ]
|
||||
#
|
||||
# ## Nodes to exclude from gathering
|
||||
# ## Nodes to include or exclude from gathering
|
||||
# ## When using both lists, node_exclude has priority.
|
||||
# # node_include = [ "*" ]
|
||||
# # node_exclude = [ ]
|
||||
#
|
||||
# ## Worker pool for jenkins plugin only
|
||||
|
|
@ -4343,19 +4518,19 @@
|
|||
# ## List of metrics collected on above servers
|
||||
# ## Each metric consists in a name, a jmx path and either
|
||||
# ## a pass or drop slice attribute.
|
||||
# ## This collect all heap memory usage metrics.
|
||||
# ## This collect all heap memory usage metrics.
|
||||
# [[inputs.jolokia.metrics]]
|
||||
# name = "heap_memory_usage"
|
||||
# mbean = "java.lang:type=Memory"
|
||||
# attribute = "HeapMemoryUsage"
|
||||
#
|
||||
# ## This collect thread counts metrics.
|
||||
# ## This collect thread counts metrics.
|
||||
# [[inputs.jolokia.metrics]]
|
||||
# name = "thread_count"
|
||||
# mbean = "java.lang:type=Threading"
|
||||
# attribute = "TotalStartedThreadCount,ThreadCount,DaemonThreadCount,PeakThreadCount"
|
||||
#
|
||||
# ## This collect number of class loaded/unloaded counts metrics.
|
||||
# ## This collect number of class loaded/unloaded counts metrics.
|
||||
# [[inputs.jolokia.metrics]]
|
||||
# name = "class_count"
|
||||
# mbean = "java.lang:type=ClassLoading"
|
||||
|
|
@ -4637,6 +4812,13 @@
|
|||
# # timeout = "5s"
|
||||
|
||||
|
||||
# # Get md array statistics from /proc/mdstat
|
||||
# [[inputs.mdstat]]
|
||||
# ## Sets file path
|
||||
# ## If not specified, then default is /proc/mdstat
|
||||
# # file_name = "/proc/mdstat"
|
||||
|
||||
|
||||
# # Read metrics from one or many memcached servers
|
||||
# [[inputs.memcached]]
|
||||
# ## An array of address to gather stats about. Specify an ip on hostname
|
||||
|
|
@ -4708,7 +4890,7 @@
|
|||
# [[inputs.modbus]]
|
||||
# ## Connection Configuration
|
||||
# ##
|
||||
# ## The plugin supports connections to PLCs via MODBUS/TCP or
|
||||
# ## The plugin supports connections to PLCs via MODBUS/TCP, RTU over TCP, ASCII over TCP or
|
||||
# ## via serial line communication in binary (RTU) or readable (ASCII) encoding
|
||||
# ##
|
||||
# ## Device name
|
||||
|
|
@ -4735,8 +4917,11 @@
|
|||
# # data_bits = 8
|
||||
# # parity = "N"
|
||||
# # stop_bits = 1
|
||||
# # transmission_mode = "RTU"
|
||||
#
|
||||
# ## For Modbus over TCP you can choose between "TCP", "RTUoverTCP" and "ASCIIoverTCP"
|
||||
# ## default behaviour is "TCP" if the controller is TCP
|
||||
# ## For Serial you can choose between "RTU" and "ASCII"
|
||||
# # transmission_mode = "RTU"
|
||||
#
|
||||
# ## Measurements
|
||||
# ##
|
||||
|
|
@ -5543,9 +5728,10 @@
|
|||
# # pattern = "nginx"
|
||||
# ## user as argument for pgrep (ie, pgrep -u <user>)
|
||||
# # user = "nginx"
|
||||
# ## Systemd unit name
|
||||
# ## Systemd unit name, supports globs when include_systemd_children is set to true
|
||||
# # systemd_unit = "nginx.service"
|
||||
# ## CGroup name or path
|
||||
# # include_systemd_children = false
|
||||
# ## CGroup name or path, supports globs
|
||||
# # cgroup = "systemd/system.slice/nginx.service"
|
||||
#
|
||||
# ## Windows service name
|
||||
|
|
@ -5785,13 +5971,6 @@
|
|||
# # Specify a list of one or more riak http servers
|
||||
# servers = ["http://localhost:8098"]
|
||||
|
||||
# # Query statistics from AMD Graphics cards using rocm-smi binary
|
||||
# [[inputs.amd_rocm_smi]]
|
||||
# ## Optional: path to rocm-smi binary, defaults to $PATH via exec.LookPath
|
||||
# # bin_path = "/opt/rocm/bin/rocm-smi"
|
||||
#
|
||||
# ## Optional: timeout for GPU polling
|
||||
# # timeout = "5s"
|
||||
|
||||
# # Read API usage and limits for a Salesforce organisation
|
||||
# [[inputs.salesforce]]
|
||||
|
|
@ -6186,6 +6365,13 @@
|
|||
# ## values are "socket", "target", "device", "mount", "automount", "swap",
|
||||
# ## "timer", "path", "slice" and "scope ":
|
||||
# # unittype = "service"
|
||||
# #
|
||||
# ## Filter for a specific pattern, default is "" (i.e. all), other possible
|
||||
# ## values are valid pattern for systemctl, e.g. "a*" for all units with
|
||||
# ## names starting with "a"
|
||||
# # pattern = ""
|
||||
# ## pattern = "telegraf* influxdb*"
|
||||
# ## pattern = "a*"
|
||||
|
||||
|
||||
# # Reads metrics from a Teamspeak 3 Server via ServerQuery
|
||||
|
|
@ -6293,6 +6479,219 @@
|
|||
# ## General connection timeout
|
||||
# # timeout = "5s"
|
||||
|
||||
# # Input plugin to collect Windows Event Log messages
|
||||
# [[inputs.win_eventlog]]
|
||||
# ## Telegraf should have Administrator permissions to subscribe for some Windows Events channels
|
||||
# ## (System log, for example)
|
||||
#
|
||||
# ## LCID (Locale ID) for event rendering
|
||||
# ## 1033 to force English language
|
||||
# ## 0 to use default Windows locale
|
||||
# # locale = 0
|
||||
#
|
||||
# ## Name of eventlog, used only if xpath_query is empty
|
||||
# ## Example: "Application"
|
||||
# # eventlog_name = ""
|
||||
#
|
||||
# ## xpath_query can be in defined short form like "Event/System[EventID=999]"
|
||||
# ## or you can form a XML Query. Refer to the Consuming Events article:
|
||||
# ## https://docs.microsoft.com/en-us/windows/win32/wes/consuming-events
|
||||
# ## XML query is the recommended form, because it is most flexible
|
||||
# ## You can create or debug XML Query by creating Custom View in Windows Event Viewer
|
||||
# ## and then copying resulting XML here
|
||||
# xpath_query = '''
|
||||
# <QueryList>
|
||||
# <Query Id="0" Path="Security">
|
||||
# <Select Path="Security">*</Select>
|
||||
# <Suppress Path="Security">*[System[( (EventID >= 5152 and EventID <= 5158) or EventID=5379 or EventID=4672)]]</Suppress>
|
||||
# </Query>
|
||||
# <Query Id="1" Path="Application">
|
||||
# <Select Path="Application">*[System[(Level < 4)]]</Select>
|
||||
# </Query>
|
||||
# <Query Id="2" Path="Windows PowerShell">
|
||||
# <Select Path="Windows PowerShell">*[System[(Level < 4)]]</Select>
|
||||
# </Query>
|
||||
# <Query Id="3" Path="System">
|
||||
# <Select Path="System">*</Select>
|
||||
# </Query>
|
||||
# <Query Id="4" Path="Setup">
|
||||
# <Select Path="Setup">*</Select>
|
||||
# </Query>
|
||||
# </QueryList>
|
||||
# '''
|
||||
#
|
||||
# ## System field names:
|
||||
# ## "Source", "EventID", "Version", "Level", "Task", "Opcode", "Keywords", "TimeCreated",
|
||||
# ## "EventRecordID", "ActivityID", "RelatedActivityID", "ProcessID", "ThreadID", "ProcessName",
|
||||
# ## "Channel", "Computer", "UserID", "UserName", "Message", "LevelText", "TaskText", "OpcodeText"
|
||||
#
|
||||
# ## In addition to System, Data fields can be unrolled from additional XML nodes in event.
|
||||
# ## Human-readable representation of those nodes is formatted into event Message field,
|
||||
# ## but XML is more machine-parsable
|
||||
#
|
||||
# # Process UserData XML to fields, if this node exists in Event XML
|
||||
# process_userdata = true
|
||||
#
|
||||
# # Process EventData XML to fields, if this node exists in Event XML
|
||||
# process_eventdata = true
|
||||
#
|
||||
# ## Separator character to use for unrolled XML Data field names
|
||||
# separator = "_"
|
||||
#
|
||||
# ## Get only first line of Message field. For most events first line is usually more than enough
|
||||
# only_first_line_of_message = true
|
||||
#
|
||||
# ## Parse timestamp from TimeCreated.SystemTime event field.
|
||||
# ## Will default to current time of telegraf processing on parsing error or if set to false
|
||||
# timestamp_from_event = true
|
||||
#
|
||||
# ## Fields to include as tags. Globbing supported ("Level*" for both "Level" and "LevelText")
|
||||
# event_tags = ["Source", "EventID", "Level", "LevelText", "Task", "TaskText", "Opcode", "OpcodeText", "Keywords", "Channel", "Computer"]
|
||||
#
|
||||
# ## Default list of fields to send. All fields are sent by default. Globbing supported
|
||||
# event_fields = ["*"]
|
||||
#
|
||||
# ## Fields to exclude. Also applied to data fields. Globbing supported
|
||||
# exclude_fields = ["TimeCreated", "Binary", "Data_Address*"]
|
||||
#
|
||||
# ## Skip those tags or fields if their value is empty or equals to zero. Globbing supported
|
||||
# exclude_empty = ["*ActivityID", "UserID"]
|
||||
|
||||
|
||||
# # Input plugin to counterPath Performance Counters on Windows operating systems
|
||||
# [[inputs.win_perf_counters]]
|
||||
# ## By default this plugin returns basic CPU and Disk statistics.
|
||||
# ## See the README file for more examples.
|
||||
# ## Uncomment examples below or write your own as you see fit. If the system
|
||||
# ## being polled for data does not have the Object at startup of the Telegraf
|
||||
# ## agent, it will not be gathered.
|
||||
# ## Settings:
|
||||
# # PrintValid = false # Print All matching performance counters
|
||||
# # Whether request a timestamp along with the PerfCounter data or just use current time
|
||||
# # UsePerfCounterTime=true
|
||||
# # If UseWildcardsExpansion params is set to true, wildcards (partial wildcards in instance names and wildcards in counters names) in configured counter paths will be expanded
|
||||
# # and in case of localized Windows, counter paths will be also localized. It also returns instance indexes in instance names.
|
||||
# # If false, wildcards (not partial) in instance names will still be expanded, but instance indexes will not be returned in instance names.
|
||||
# #UseWildcardsExpansion = false
|
||||
# # Period after which counters will be reread from configuration and wildcards in counter paths expanded
|
||||
# CountersRefreshInterval="1m"
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# # Processor usage, alternative to native, reports on a per core.
|
||||
# ObjectName = "Processor"
|
||||
# Instances = ["*"]
|
||||
# Counters = [
|
||||
# "% Idle Time",
|
||||
# "% Interrupt Time",
|
||||
# "% Privileged Time",
|
||||
# "% User Time",
|
||||
# "% Processor Time",
|
||||
# "% DPC Time",
|
||||
# ]
|
||||
# Measurement = "win_cpu"
|
||||
# # Set to true to include _Total instance when querying for all (*).
|
||||
# # IncludeTotal=false
|
||||
# # Print out when the performance counter is missing from object, counter or instance.
|
||||
# # WarnOnMissing = false
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# # Disk times and queues
|
||||
# ObjectName = "LogicalDisk"
|
||||
# Instances = ["*"]
|
||||
# Counters = [
|
||||
# "% Idle Time",
|
||||
# "% Disk Time",
|
||||
# "% Disk Read Time",
|
||||
# "% Disk Write Time",
|
||||
# "% User Time",
|
||||
# "% Free Space",
|
||||
# "Current Disk Queue Length",
|
||||
# "Free Megabytes",
|
||||
# ]
|
||||
# Measurement = "win_disk"
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# ObjectName = "PhysicalDisk"
|
||||
# Instances = ["*"]
|
||||
# Counters = [
|
||||
# "Disk Read Bytes/sec",
|
||||
# "Disk Write Bytes/sec",
|
||||
# "Current Disk Queue Length",
|
||||
# "Disk Reads/sec",
|
||||
# "Disk Writes/sec",
|
||||
# "% Disk Time",
|
||||
# "% Disk Read Time",
|
||||
# "% Disk Write Time",
|
||||
# ]
|
||||
# Measurement = "win_diskio"
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# ObjectName = "Network Interface"
|
||||
# Instances = ["*"]
|
||||
# Counters = [
|
||||
# "Bytes Received/sec",
|
||||
# "Bytes Sent/sec",
|
||||
# "Packets Received/sec",
|
||||
# "Packets Sent/sec",
|
||||
# "Packets Received Discarded",
|
||||
# "Packets Outbound Discarded",
|
||||
# "Packets Received Errors",
|
||||
# "Packets Outbound Errors",
|
||||
# ]
|
||||
# Measurement = "win_net"
|
||||
#
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# ObjectName = "System"
|
||||
# Counters = [
|
||||
# "Context Switches/sec",
|
||||
# "System Calls/sec",
|
||||
# "Processor Queue Length",
|
||||
# "System Up Time",
|
||||
# ]
|
||||
# Instances = ["------"]
|
||||
# Measurement = "win_system"
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# # Example counterPath where the Instance portion must be removed to get data back,
|
||||
# # such as from the Memory object.
|
||||
# ObjectName = "Memory"
|
||||
# Counters = [
|
||||
# "Available Bytes",
|
||||
# "Cache Faults/sec",
|
||||
# "Demand Zero Faults/sec",
|
||||
# "Page Faults/sec",
|
||||
# "Pages/sec",
|
||||
# "Transition Faults/sec",
|
||||
# "Pool Nonpaged Bytes",
|
||||
# "Pool Paged Bytes",
|
||||
# "Standby Cache Reserve Bytes",
|
||||
# "Standby Cache Normal Priority Bytes",
|
||||
# "Standby Cache Core Bytes",
|
||||
# ]
|
||||
# Instances = ["------"] # Use 6 x - to remove the Instance bit from the counterPath.
|
||||
# Measurement = "win_mem"
|
||||
#
|
||||
# [[inputs.win_perf_counters.object]]
|
||||
# # Example query where the Instance portion must be removed to get data back,
|
||||
# # such as from the Paging File object.
|
||||
# ObjectName = "Paging File"
|
||||
# Counters = [
|
||||
# "% Usage",
|
||||
# ]
|
||||
# Instances = ["_Total"]
|
||||
# Measurement = "win_swap"
|
||||
|
||||
|
||||
# # Input plugin to report Windows services info.
|
||||
# [[inputs.win_services]]
|
||||
# ## Names of the services to monitor. Leave empty to monitor all the available services on the host. Globs accepted.
|
||||
# service_names = [
|
||||
# "LanmanServer",
|
||||
# "TermService",
|
||||
# "Win*",
|
||||
# ]
|
||||
|
||||
|
||||
# # A plugin to collect stats from Varnish HTTP Cache
|
||||
# [[inputs.varnish]]
|
||||
|
|
@ -7138,7 +7537,14 @@
|
|||
# service_address = ":8080"
|
||||
#
|
||||
# ## Path to listen to.
|
||||
# # path = "/telegraf"
|
||||
# ## This option is deprecated and only available for backward-compatibility. Please use paths instead.
|
||||
# # path = ""
|
||||
#
|
||||
# ## Paths to listen to.
|
||||
# # paths = ["/telegraf"]
|
||||
#
|
||||
# ## Save path as http_listener_v2_path tag if set to true
|
||||
# # path_tag = false
|
||||
#
|
||||
# ## HTTP methods to accept.
|
||||
# # methods = ["POST", "PUT"]
|
||||
|
|
@ -7149,7 +7555,7 @@
|
|||
# # write_timeout = "10s"
|
||||
#
|
||||
# ## Maximum allowed http request body size in bytes.
|
||||
# ## 0 means to use the default of 524,288,00 bytes (500 mebibytes)
|
||||
# ## 0 means to use the default of 524,288,000 bytes (500 mebibytes)
|
||||
# # max_body_size = "500MB"
|
||||
#
|
||||
# ## Part of the request to consume. Available options are "body" and
|
||||
|
|
@ -7450,16 +7856,19 @@
|
|||
#
|
||||
# ## Amazon Credentials
|
||||
# ## Credentials are loaded in the following order
|
||||
# ## 1) Assumed credentials via STS if role_arn is specified
|
||||
# ## 2) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 3) shared profile from 'profile'
|
||||
# ## 4) environment variables
|
||||
# ## 5) shared credentials file
|
||||
# ## 6) EC2 Instance Profile
|
||||
# ## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||
# ## 2) Assumed credentials via STS if role_arn is specified
|
||||
# ## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||
# ## 4) shared profile from 'profile'
|
||||
# ## 5) environment variables
|
||||
# ## 6) shared credentials file
|
||||
# ## 7) EC2 Instance Profile
|
||||
# # access_key = ""
|
||||
# # secret_key = ""
|
||||
# # token = ""
|
||||
# # role_arn = ""
|
||||
# # web_identity_token_file = ""
|
||||
# # role_session_name = ""
|
||||
# # profile = ""
|
||||
# # shared_credential_file = ""
|
||||
#
|
||||
|
|
@ -7723,16 +8132,28 @@
|
|||
|
||||
# # Receive OpenTelemetry traces, metrics, and logs over gRPC
|
||||
# [[inputs.opentelemetry]]
|
||||
# ## Override the OpenTelemetry gRPC service address:port
|
||||
# ## Override the default (0.0.0.0:4317) destination OpenTelemetry gRPC service
|
||||
# ## address:port
|
||||
# # service_address = "0.0.0.0:4317"
|
||||
#
|
||||
# ## Override the default request timeout
|
||||
# ## Override the default (5s) new connection timeout
|
||||
# # timeout = "5s"
|
||||
#
|
||||
# ## Select a schema for metrics: prometheus-v1 or prometheus-v2
|
||||
# ## Override the default (prometheus-v1) metrics schema.
|
||||
# ## Supports: "prometheus-v1", "prometheus-v2"
|
||||
# ## For more information about the alternatives, read the Prometheus input
|
||||
# ## plugin notes.
|
||||
# # metrics_schema = "prometheus-v1"
|
||||
#
|
||||
# ## Optional TLS Config.
|
||||
# ## For advanced options: https://github.com/influxdata/telegraf/blob/v1.18.3/docs/TLS.md
|
||||
# ##
|
||||
# ## Set one or more allowed client CA certificate file names to
|
||||
# ## enable mutually authenticated TLS connections.
|
||||
# # tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]
|
||||
# ## Add service certificate and key.
|
||||
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||
# # tls_key = "/etc/telegraf/key.pem"
|
||||
|
||||
|
||||
# # Read metrics from one or many pgbouncer servers
|
||||
|
|
@ -7909,6 +8330,19 @@
|
|||
# # eg. To scrape pods on a specific node
|
||||
# # kubernetes_field_selector = "spec.nodeName=$HOSTNAME"
|
||||
#
|
||||
# ## Scrape Services available in Consul Catalog
|
||||
# # [inputs.prometheus.consul]
|
||||
# # enabled = true
|
||||
# # agent = "http://localhost:8500"
|
||||
# # query_interval = "5m"
|
||||
#
|
||||
# # [[inputs.prometheus.consul.query]]
|
||||
# # name = "a service name"
|
||||
# # tag = "a service tag"
|
||||
# # url = 'http://{{if ne .ServiceAddress ""}}{{.ServiceAddress}}{{else}}{{.Address}}{{end}}:{{.ServicePort}}/{{with .ServiceMeta.metrics_path}}{{.}}{{else}}metrics{{end}}'
|
||||
# # [inputs.prometheus.consul.query.tags]
|
||||
# # host = "{{.Node}}"
|
||||
#
|
||||
# ## Use bearer token for authorization. ('bearer_token' takes priority)
|
||||
# # bearer_token = "/path/to/bearer/token"
|
||||
# ## OR
|
||||
|
|
@ -7989,6 +8423,10 @@
|
|||
# ## 1024. See README.md for details
|
||||
# ##
|
||||
# # service_address = "udp://:162"
|
||||
# ##
|
||||
# ## Path to mib files
|
||||
# # path = ["/usr/share/snmp/mibs"]
|
||||
# ##
|
||||
# ## Timeout running snmptranslate command
|
||||
# # timeout = "5s"
|
||||
# ## Snmp version, defaults to 2c
|
||||
|
|
@ -8082,6 +8520,7 @@
|
|||
# dsn = "username:password@mysqlserver:3307/dbname?param=value"
|
||||
#
|
||||
# ## Timeout for any operation
|
||||
# ## Note that the timeout for queries is per query not per gather.
|
||||
# # timeout = "5s"
|
||||
#
|
||||
# ## Connection time limits
|
||||
|
|
@ -8282,9 +8721,9 @@
|
|||
# #max_ttl = "1000h"
|
||||
|
||||
|
||||
# # Suricata stats plugin
|
||||
# # Suricata stats and alerts plugin
|
||||
# [[inputs.suricata]]
|
||||
# ## Data sink for Suricata stats log
|
||||
# ## Data sink for Suricata stats and alerts logs
|
||||
# # This is expected to be a filename of a
|
||||
# # unix socket to be created for listening.
|
||||
# source = "/var/run/suricata-stats.sock"
|
||||
|
|
@ -8292,6 +8731,9 @@
|
|||
# # Delimiter for flattening field keys, e.g. subitem "alert" of "detect"
|
||||
# # becomes "detect_alert" when delimiter is "_".
|
||||
# delimiter = "_"
|
||||
#
|
||||
# ## Detect alert logs
|
||||
# # alerts = false
|
||||
|
||||
|
||||
# # Accepts syslog messages following RFC5424 format with transports as per RFC5426, RFC5425, or RFC6587
|
||||
|
|
@ -8336,6 +8778,11 @@
|
|||
# ## By default best effort parsing is off.
|
||||
# # best_effort = false
|
||||
#
|
||||
# ## The RFC standard to use for message parsing
|
||||
# ## By default RFC5424 is used. RFC3164 only supports UDP transport (no streaming support)
|
||||
# ## Must be one of "RFC5424", or "RFC3164".
|
||||
# # syslog_standard = "RFC5424"
|
||||
#
|
||||
# ## Character to prepend to SD-PARAMs (default = "_").
|
||||
# ## A syslog message can contain multiple parameters and multiple identifiers within structured data section.
|
||||
# ## Eg., [id1 name1="val1" name2="val2"][id2 name1="val1" nameA="valA"]
|
||||
|
|
|
|||
Loading…
Reference in New Issue