docs: Update etc/telegraf.conf and etc/telegraf_windows.conf (#11365)

Co-authored-by: Tiger Bot <>
This commit is contained in:
telegraf-tiger[bot] 2022-06-22 15:12:18 -05:00 committed by GitHub
parent 3f114e0921
commit d91ec1466b
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 612 additions and 258 deletions

View File

@ -21,96 +21,96 @@
## Environment variables can be used as tags, and throughout the config file
# user = "$USER"
# Configuration for telegraf agent
[agent]
## Default data collection interval for all inputs
interval = "10s"
## Rounds collection interval to 'interval'
## ie, if interval="10s" then always collect on :00, :10, :20, etc.
round_interval = true
## Telegraf will send metrics to outputs in batches of at most
## metric_batch_size metrics.
## This controls the size of writes that Telegraf sends to output plugins.
metric_batch_size = 1000
## Maximum number of unwritten metrics per output. Increasing this value
## allows for longer periods of output downtime without dropping metrics at the
## cost of higher maximum memory usage.
metric_buffer_limit = 10000
## Collection jitter is used to jitter the collection by a random amount.
## Each plugin will sleep for a random time within jitter before collecting.
## This can be used to avoid many plugins querying things like sysfs at the
## same time, which can have a measurable effect on the system.
collection_jitter = "0s"
## Collection offset is used to shift the collection by the given amount.
## This can be be used to avoid many plugins querying constraint devices
## at the same time by manually scheduling them in time.
# collection_offset = "0s"
## Default flushing interval for all outputs. Maximum flush_interval will be
## flush_interval + flush_jitter
flush_interval = "10s"
## Jitter the flush interval by a random amount. This is primarily to avoid
## large write spikes for users running a large number of telegraf instances.
## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s
flush_jitter = "0s"
## Collected metrics are rounded to the precision specified. Precision is
## specified as an interval with an integer + unit (e.g. 0s, 10ms, 2us, 4s).
## Valid time units are "ns", "us" (or "µs"), "ms", "s".
##
## By default or when set to "0s", precision will be set to the same
## timestamp order as the collection interval, with the maximum being 1s:
## ie, when interval = "10s", precision will be "1s"
## when interval = "250ms", precision will be "1ms"
##
## Precision will NOT be used for service inputs. It is up to each individual
## service input to set the timestamp at the appropriate precision.
precision = "0s"
## Log at debug level.
# debug = false
## Log only error level messages.
# quiet = false
## Log target controls the destination for logs and can be one of "file",
## "stderr" or, on Windows, "eventlog". When set to "file", the output file
## is determined by the "logfile" setting.
# logtarget = "file"
## Name of the file to be logged to when using the "file" logtarget. If set to
## the empty string then logs are written to stderr.
# logfile = ""
## The logfile will be rotated after the time interval specified. When set
## to 0 no time based rotation is performed. Logs are rotated only when
## written to, if there is no log activity rotation may be delayed.
# logfile_rotation_interval = "0h"
## The logfile will be rotated when it becomes larger than the specified
## size. When set to 0 no size based rotation is performed.
# logfile_rotation_max_size = "0MB"
## Maximum number of rotated archives to keep, any older logs are deleted.
## If set to -1, no archives are removed.
# logfile_rotation_max_archives = 5
## Pick a timezone to use when logging or type 'local' for local time.
## Example: America/Chicago
# log_with_timezone = ""
## Override default hostname, if empty use os.Hostname()
hostname = ""
## If set to true, do no set the "host" tag in the telegraf agent.
omit_hostname = false
## Method of translating SNMP objects. Can be "netsnmp" which
## translates by calling external programs snmptranslate and snmptable,
## or "gosmi" which translates using the built-in gosmi library.
# snmp_translator = "netsnmp"
# Configuration for telegraf agent
[agent]
## Default data collection interval for all inputs
interval = "10s"
## Rounds collection interval to 'interval'
## ie, if interval="10s" then always collect on :00, :10, :20, etc.
round_interval = true
## Telegraf will send metrics to outputs in batches of at most
## metric_batch_size metrics.
## This controls the size of writes that Telegraf sends to output plugins.
metric_batch_size = 1000
## Maximum number of unwritten metrics per output. Increasing this value
## allows for longer periods of output downtime without dropping metrics at the
## cost of higher maximum memory usage.
metric_buffer_limit = 10000
## Collection jitter is used to jitter the collection by a random amount.
## Each plugin will sleep for a random time within jitter before collecting.
## This can be used to avoid many plugins querying things like sysfs at the
## same time, which can have a measurable effect on the system.
collection_jitter = "0s"
## Collection offset is used to shift the collection by the given amount.
## This can be be used to avoid many plugins querying constraint devices
## at the same time by manually scheduling them in time.
# collection_offset = "0s"
## Default flushing interval for all outputs. Maximum flush_interval will be
## flush_interval + flush_jitter
flush_interval = "10s"
## Jitter the flush interval by a random amount. This is primarily to avoid
## large write spikes for users running a large number of telegraf instances.
## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s
flush_jitter = "0s"
## Collected metrics are rounded to the precision specified. Precision is
## specified as an interval with an integer + unit (e.g. 0s, 10ms, 2us, 4s).
## Valid time units are "ns", "us" (or "µs"), "ms", "s".
##
## By default or when set to "0s", precision will be set to the same
## timestamp order as the collection interval, with the maximum being 1s:
## ie, when interval = "10s", precision will be "1s"
## when interval = "250ms", precision will be "1ms"
##
## Precision will NOT be used for service inputs. It is up to each individual
## service input to set the timestamp at the appropriate precision.
precision = "0s"
## Log at debug level.
# debug = false
## Log only error level messages.
# quiet = false
## Log target controls the destination for logs and can be one of "file",
## "stderr" or, on Windows, "eventlog". When set to "file", the output file
## is determined by the "logfile" setting.
# logtarget = "file"
## Name of the file to be logged to when using the "file" logtarget. If set to
## the empty string then logs are written to stderr.
# logfile = ""
## The logfile will be rotated after the time interval specified. When set
## to 0 no time based rotation is performed. Logs are rotated only when
## written to, if there is no log activity rotation may be delayed.
# logfile_rotation_interval = "0h"
## The logfile will be rotated when it becomes larger than the specified
## size. When set to 0 no size based rotation is performed.
# logfile_rotation_max_size = "0MB"
## Maximum number of rotated archives to keep, any older logs are deleted.
## If set to -1, no archives are removed.
# logfile_rotation_max_archives = 5
## Pick a timezone to use when logging or type 'local' for local time.
## Example: America/Chicago
# log_with_timezone = ""
## Override default hostname, if empty use os.Hostname()
hostname = ""
## If set to true, do no set the "host" tag in the telegraf agent.
omit_hostname = false
## Method of translating SNMP objects. Can be "netsnmp" which
## translates by calling external programs snmptranslate and snmptable,
## or "gosmi" which translates using the built-in gosmi library.
# snmp_translator = "netsnmp"
###############################################################################
# OUTPUT PLUGINS #
@ -1353,24 +1353,16 @@
# # A plugin that can send metrics over HTTPs to Logz.io
# [[outputs.logzio]]
# ## Set to true if Logz.io sender checks the disk space before adding metrics to the disk queue.
# # check_disk_space = true
# ## Connection timeout, defaults to "5s" if not set.
# # timeout = "5s"
#
# ## The percent of used file system space at which the sender will stop queueing.
# ## When we will reach that percentage, the file system in which the queue is stored will drop
# ## all new logs until the percentage of used space drops below that threshold.
# # disk_threshold = 98
#
# ## How often Logz.io sender should drain the queue.
# ## Valid time units are "ns", "us" (or "µs"), "ms", "s", "m", "h".
# # drain_duration = "3s"
#
# ## Where Logz.io sender should store the queue
# ## queue_dir = Sprintf("%s%s%s%s%d", os.TempDir(), string(os.PathSeparator),
# ## "logzio-buffer", string(os.PathSeparator), time.Now().UnixNano())
# ## Optional TLS Config
# # tls_ca = "/etc/telegraf/ca.pem"
# # tls_cert = "/etc/telegraf/cert.pem"
# # tls_key = "/etc/telegraf/key.pem"
#
# ## Logz.io account token
# token = "your Logz.io token" # required
# token = "your logz.io token" # required
#
# ## Use your listener URL for your Logz.io account region.
# # url = "https://listener.logz.io:8071"
@ -3036,21 +3028,20 @@
# # Calculates a derivative for every field.
# [[aggregators.derivative]]
# ## Specific Derivative Aggregator Arguments:
#
# ## Configure a custom derivation variable. Timestamp is used if none is given.
# # variable = ""
#
# ## Suffix to add to the field name for the derivative name.
# # suffix = "_rate"
#
# ## Roll-Over last measurement to first measurement of next period
# # max_roll_over = 10
#
# ## General Aggregator Arguments:
#
# ## calculate derivative every 30 seconds
# ## The period in which to flush the aggregator.
# period = "30s"
# ##
# ## Suffix to append for the resulting derivative field.
# # suffix = "_rate"
# ##
# ## Field to use for the quotient when computing the derivative.
# ## When using a field as the derivation parameter the name of that field will
# ## be used for the resulting derivative, e.g. *fieldname_by_parameter*.
# ## By default the timestamps of the metrics are used and the suffix is omitted.
# # variable = ""
# ##
# ## Maximum number of roll-overs in case only one measurement is found during a period.
# # max_roll_over = 10
# # Report the final metric of a series
@ -6853,7 +6844,6 @@
# ## DEPRECATED: The 'snmp_legacy' plugin is deprecated in version 1.0.0, use 'inputs.snmp' instead.
# # DEPRECATED! PLEASE USE inputs.snmp INSTEAD.
# [[inputs.snmp_legacy]]
# ## Use 'oids.txt' file to translate oids to names
# ## To generate 'oids.txt' you need to run:
@ -7587,11 +7577,6 @@
# ## DEPRECATED: The 'cassandra' plugin is deprecated in version 1.7.0, use 'inputs.jolokia2' with the 'cassandra.conf' example configuration instead.
# # Read Cassandra metrics through Jolokia
# [[inputs.cassandra]]
# ## DEPRECATED: The cassandra plugin has been deprecated. Please use the
# ## jolokia2 plugin instead.
# ##
# ## see https://github.com/influxdata/telegraf/tree/master/plugins/inputs/jolokia2
#
# context = "/jolokia/read"
# ## List of cassandra servers exposing jolokia read service
# servers = ["myuser:mypassword@10.10.10.1:8778","10.10.10.2:8778",":8778"]
@ -8537,7 +8522,6 @@
# ## DEPRECATED: The 'kafka_consumer_legacy' plugin is deprecated in version 1.4.0, use 'inputs.kafka_consumer' instead, NOTE: 'kafka_consumer' only supports Kafka v0.8+.
# ## DEPRECATED: The 'kafka_consumer_legacy' plugin is deprecated in version 1.4.0, use 'inputs.kafka_consumer' instead, NOTE: 'kafka_consumer' only supports Kafka v0.8+.
# # Read metrics from Kafka topic(s)
# [[inputs.kafka_consumer_legacy]]
# ## topic(s) to consume
@ -8668,7 +8652,6 @@
# ## DEPRECATED: The 'logparser' plugin is deprecated in version 1.15.0, use 'inputs.tail' with 'grok' data format instead.
# ## DEPRECATED: The 'logparser' plugin is deprecated in version 1.15.0, use 'inputs.tail' with 'grok' data format instead.
# # Read metrics off Arista LANZ, via socket
# [[inputs.logparser]]
# ## Log files to parse.
@ -9156,7 +9139,7 @@
# # Riemann protobuff listener
# [[inputs.rimann_listener]]
# [[inputs.riemann_listener]]
# ## URL to listen on
# ## Default is "tcp://:5555"
# # service_address = "tcp://:8094"
@ -9749,7 +9732,201 @@
# # see https://github.com/influxdata/telegraf/tree/master/plugins/inputs/socket_listener
# vm_metric_exclude = [ "*" ]
# -# Read metrics from one or many vCenters
# [[inputs.vsphere]]
# ## List of vCenter URLs to be monitored. These three lines must be uncommented
# ## and edited for the plugin to work.
# vcenters = [ "https://vcenter.local/sdk" ]
# username = "user@corp.local"
# password = "secret"
#
# ## VMs
# ## Typical VM metrics (if omitted or empty, all metrics are collected)
# # vm_include = [ "/*/vm/**"] # Inventory path to VMs to collect (by default all are collected)
# # vm_exclude = [] # Inventory paths to exclude
# vm_metric_include = [
# "cpu.demand.average",
# "cpu.idle.summation",
# "cpu.latency.average",
# "cpu.readiness.average",
# "cpu.ready.summation",
# "cpu.run.summation",
# "cpu.usagemhz.average",
# "cpu.used.summation",
# "cpu.wait.summation",
# "mem.active.average",
# "mem.granted.average",
# "mem.latency.average",
# "mem.swapin.average",
# "mem.swapinRate.average",
# "mem.swapout.average",
# "mem.swapoutRate.average",
# "mem.usage.average",
# "mem.vmmemctl.average",
# "net.bytesRx.average",
# "net.bytesTx.average",
# "net.droppedRx.summation",
# "net.droppedTx.summation",
# "net.usage.average",
# "power.power.average",
# "virtualDisk.numberReadAveraged.average",
# "virtualDisk.numberWriteAveraged.average",
# "virtualDisk.read.average",
# "virtualDisk.readOIO.latest",
# "virtualDisk.throughput.usage.average",
# "virtualDisk.totalReadLatency.average",
# "virtualDisk.totalWriteLatency.average",
# "virtualDisk.write.average",
# "virtualDisk.writeOIO.latest",
# "sys.uptime.latest",
# ]
# # vm_metric_exclude = [] ## Nothing is excluded by default
# # vm_instances = true ## true by default
#
# ## Hosts
# ## Typical host metrics (if omitted or empty, all metrics are collected)
# # host_include = [ "/*/host/**"] # Inventory path to hosts to collect (by default all are collected)
# # host_exclude [] # Inventory paths to exclude
# host_metric_include = [
# "cpu.coreUtilization.average",
# "cpu.costop.summation",
# "cpu.demand.average",
# "cpu.idle.summation",
# "cpu.latency.average",
# "cpu.readiness.average",
# "cpu.ready.summation",
# "cpu.swapwait.summation",
# "cpu.usage.average",
# "cpu.usagemhz.average",
# "cpu.used.summation",
# "cpu.utilization.average",
# "cpu.wait.summation",
# "disk.deviceReadLatency.average",
# "disk.deviceWriteLatency.average",
# "disk.kernelReadLatency.average",
# "disk.kernelWriteLatency.average",
# "disk.numberReadAveraged.average",
# "disk.numberWriteAveraged.average",
# "disk.read.average",
# "disk.totalReadLatency.average",
# "disk.totalWriteLatency.average",
# "disk.write.average",
# "mem.active.average",
# "mem.latency.average",
# "mem.state.latest",
# "mem.swapin.average",
# "mem.swapinRate.average",
# "mem.swapout.average",
# "mem.swapoutRate.average",
# "mem.totalCapacity.average",
# "mem.usage.average",
# "mem.vmmemctl.average",
# "net.bytesRx.average",
# "net.bytesTx.average",
# "net.droppedRx.summation",
# "net.droppedTx.summation",
# "net.errorsRx.summation",
# "net.errorsTx.summation",
# "net.usage.average",
# "power.power.average",
# "storageAdapter.numberReadAveraged.average",
# "storageAdapter.numberWriteAveraged.average",
# "storageAdapter.read.average",
# "storageAdapter.write.average",
# "sys.uptime.latest",
# ]
# ## Collect IP addresses? Valid values are "ipv4" and "ipv6"
# # ip_addresses = ["ipv6", "ipv4" ]
#
# # host_metric_exclude = [] ## Nothing excluded by default
# # host_instances = true ## true by default
#
#
# ## Clusters
# # cluster_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
# # cluster_exclude = [] # Inventory paths to exclude
# # cluster_metric_include = [] ## if omitted or empty, all metrics are collected
# # cluster_metric_exclude = [] ## Nothing excluded by default
# # cluster_instances = false ## false by default
#
# ## Resource Pools
# # datastore_include = [ "/*/host/**"] # Inventory path to datastores to collect (by default all are collected)
# # datastore_exclude = [] # Inventory paths to exclude
# # datastore_metric_include = [] ## if omitted or empty, all metrics are collected
# # datastore_metric_exclude = [] ## Nothing excluded by default
# # datastore_instances = false ## false by default
#
# ## Datastores
# # datastore_include = [ "/*/datastore/**"] # Inventory path to datastores to collect (by default all are collected)
# # datastore_exclude = [] # Inventory paths to exclude
# # datastore_metric_include = [] ## if omitted or empty, all metrics are collected
# # datastore_metric_exclude = [] ## Nothing excluded by default
# # datastore_instances = false ## false by default
#
# ## Datacenters
# # datacenter_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
# # datacenter_exclude = [] # Inventory paths to exclude
# datacenter_metric_include = [] ## if omitted or empty, all metrics are collected
# datacenter_metric_exclude = [ "*" ] ## Datacenters are not collected by default.
# # datacenter_instances = false ## false by default
#
# ## Plugin Settings
# ## separator character to use for measurement and field names (default: "_")
# # separator = "_"
#
# ## number of objects to retrieve per query for realtime resources (vms and hosts)
# ## set to 64 for vCenter 5.5 and 6.0 (default: 256)
# # max_query_objects = 256
#
# ## number of metrics to retrieve per query for non-realtime resources (clusters and datastores)
# ## set to 64 for vCenter 5.5 and 6.0 (default: 256)
# # max_query_metrics = 256
#
# ## number of go routines to use for collection and discovery of objects and metrics
# # collect_concurrency = 1
# # discover_concurrency = 1
#
# ## the interval before (re)discovering objects subject to metrics collection (default: 300s)
# # object_discovery_interval = "300s"
#
# ## timeout applies to any of the api request made to vcenter
# # timeout = "60s"
#
# ## When set to true, all samples are sent as integers. This makes the output
# ## data types backwards compatible with Telegraf 1.9 or lower. Normally all
# ## samples from vCenter, with the exception of percentages, are integer
# ## values, but under some conditions, some averaging takes place internally in
# ## the plugin. Setting this flag to "false" will send values as floats to
# ## preserve the full precision when averaging takes place.
# # use_int_samples = true
#
# ## Custom attributes from vCenter can be very useful for queries in order to slice the
# ## metrics along different dimension and for forming ad-hoc relationships. They are disabled
# ## by default, since they can add a considerable amount of tags to the resulting metrics. To
# ## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
# ## to select the attributes you want to include.
# ## By default, since they can add a considerable amount of tags to the resulting metrics. To
# ## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
# ## to select the attributes you want to include.
# # custom_attribute_include = []
# # custom_attribute_exclude = ["*"]
#
# ## The number of vSphere 5 minute metric collection cycles to look back for non-realtime metrics. In
# ## some versions (6.7, 7.0 and possible more), certain metrics, such as cluster metrics, may be reported
# ## with a significant delay (>30min). If this happens, try increasing this number. Please note that increasing
# ## it too much may cause performance issues.
# # metric_lookback = 3
#
# ## Optional SSL Config
# # ssl_ca = "/path/to/cafile"
# # ssl_cert = "/path/to/certfile"
# # ssl_key = "/path/to/keyfile"
# ## Use SSL but skip chain & host verification
# # insecure_skip_verify = false
#
# ## The Historical Interval value must match EXACTLY the interval in the daily
# # "Interval Duration" found on the VCenter server under Configure > General > Statistics > Statistic intervals
# # historical_interval = "5m"
# # A Webhooks Event collector

View File

@ -21,96 +21,96 @@
## Environment variables can be used as tags, and throughout the config file
# user = "$USER"
# Configuration for telegraf agent
[agent]
## Default data collection interval for all inputs
interval = "10s"
## Rounds collection interval to 'interval'
## ie, if interval="10s" then always collect on :00, :10, :20, etc.
round_interval = true
## Telegraf will send metrics to outputs in batches of at most
## metric_batch_size metrics.
## This controls the size of writes that Telegraf sends to output plugins.
metric_batch_size = 1000
## Maximum number of unwritten metrics per output. Increasing this value
## allows for longer periods of output downtime without dropping metrics at the
## cost of higher maximum memory usage.
metric_buffer_limit = 10000
## Collection jitter is used to jitter the collection by a random amount.
## Each plugin will sleep for a random time within jitter before collecting.
## This can be used to avoid many plugins querying things like sysfs at the
## same time, which can have a measurable effect on the system.
collection_jitter = "0s"
## Collection offset is used to shift the collection by the given amount.
## This can be be used to avoid many plugins querying constraint devices
## at the same time by manually scheduling them in time.
# collection_offset = "0s"
## Default flushing interval for all outputs. Maximum flush_interval will be
## flush_interval + flush_jitter
flush_interval = "10s"
## Jitter the flush interval by a random amount. This is primarily to avoid
## large write spikes for users running a large number of telegraf instances.
## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s
flush_jitter = "0s"
## Collected metrics are rounded to the precision specified. Precision is
## specified as an interval with an integer + unit (e.g. 0s, 10ms, 2us, 4s).
## Valid time units are "ns", "us" (or "µs"), "ms", "s".
##
## By default or when set to "0s", precision will be set to the same
## timestamp order as the collection interval, with the maximum being 1s:
## ie, when interval = "10s", precision will be "1s"
## when interval = "250ms", precision will be "1ms"
##
## Precision will NOT be used for service inputs. It is up to each individual
## service input to set the timestamp at the appropriate precision.
precision = "0s"
## Log at debug level.
# debug = false
## Log only error level messages.
# quiet = false
## Log target controls the destination for logs and can be one of "file",
## "stderr" or, on Windows, "eventlog". When set to "file", the output file
## is determined by the "logfile" setting.
# logtarget = "file"
## Name of the file to be logged to when using the "file" logtarget. If set to
## the empty string then logs are written to stderr.
# logfile = ""
## The logfile will be rotated after the time interval specified. When set
## to 0 no time based rotation is performed. Logs are rotated only when
## written to, if there is no log activity rotation may be delayed.
# logfile_rotation_interval = "0h"
## The logfile will be rotated when it becomes larger than the specified
## size. When set to 0 no size based rotation is performed.
# logfile_rotation_max_size = "0MB"
## Maximum number of rotated archives to keep, any older logs are deleted.
## If set to -1, no archives are removed.
# logfile_rotation_max_archives = 5
## Pick a timezone to use when logging or type 'local' for local time.
## Example: America/Chicago
# log_with_timezone = ""
## Override default hostname, if empty use os.Hostname()
hostname = ""
## If set to true, do no set the "host" tag in the telegraf agent.
omit_hostname = false
## Method of translating SNMP objects. Can be "netsnmp" which
## translates by calling external programs snmptranslate and snmptable,
## or "gosmi" which translates using the built-in gosmi library.
# snmp_translator = "netsnmp"
# Configuration for telegraf agent
[agent]
## Default data collection interval for all inputs
interval = "10s"
## Rounds collection interval to 'interval'
## ie, if interval="10s" then always collect on :00, :10, :20, etc.
round_interval = true
## Telegraf will send metrics to outputs in batches of at most
## metric_batch_size metrics.
## This controls the size of writes that Telegraf sends to output plugins.
metric_batch_size = 1000
## Maximum number of unwritten metrics per output. Increasing this value
## allows for longer periods of output downtime without dropping metrics at the
## cost of higher maximum memory usage.
metric_buffer_limit = 10000
## Collection jitter is used to jitter the collection by a random amount.
## Each plugin will sleep for a random time within jitter before collecting.
## This can be used to avoid many plugins querying things like sysfs at the
## same time, which can have a measurable effect on the system.
collection_jitter = "0s"
## Collection offset is used to shift the collection by the given amount.
## This can be be used to avoid many plugins querying constraint devices
## at the same time by manually scheduling them in time.
# collection_offset = "0s"
## Default flushing interval for all outputs. Maximum flush_interval will be
## flush_interval + flush_jitter
flush_interval = "10s"
## Jitter the flush interval by a random amount. This is primarily to avoid
## large write spikes for users running a large number of telegraf instances.
## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s
flush_jitter = "0s"
## Collected metrics are rounded to the precision specified. Precision is
## specified as an interval with an integer + unit (e.g. 0s, 10ms, 2us, 4s).
## Valid time units are "ns", "us" (or "µs"), "ms", "s".
##
## By default or when set to "0s", precision will be set to the same
## timestamp order as the collection interval, with the maximum being 1s:
## ie, when interval = "10s", precision will be "1s"
## when interval = "250ms", precision will be "1ms"
##
## Precision will NOT be used for service inputs. It is up to each individual
## service input to set the timestamp at the appropriate precision.
precision = "0s"
## Log at debug level.
# debug = false
## Log only error level messages.
# quiet = false
## Log target controls the destination for logs and can be one of "file",
## "stderr" or, on Windows, "eventlog". When set to "file", the output file
## is determined by the "logfile" setting.
# logtarget = "file"
## Name of the file to be logged to when using the "file" logtarget. If set to
## the empty string then logs are written to stderr.
# logfile = ""
## The logfile will be rotated after the time interval specified. When set
## to 0 no time based rotation is performed. Logs are rotated only when
## written to, if there is no log activity rotation may be delayed.
# logfile_rotation_interval = "0h"
## The logfile will be rotated when it becomes larger than the specified
## size. When set to 0 no size based rotation is performed.
# logfile_rotation_max_size = "0MB"
## Maximum number of rotated archives to keep, any older logs are deleted.
## If set to -1, no archives are removed.
# logfile_rotation_max_archives = 5
## Pick a timezone to use when logging or type 'local' for local time.
## Example: America/Chicago
# log_with_timezone = ""
## Override default hostname, if empty use os.Hostname()
hostname = ""
## If set to true, do no set the "host" tag in the telegraf agent.
omit_hostname = false
## Method of translating SNMP objects. Can be "netsnmp" which
## translates by calling external programs snmptranslate and snmptable,
## or "gosmi" which translates using the built-in gosmi library.
# snmp_translator = "netsnmp"
###############################################################################
# OUTPUT PLUGINS #
@ -1353,24 +1353,16 @@
# # A plugin that can send metrics over HTTPs to Logz.io
# [[outputs.logzio]]
# ## Set to true if Logz.io sender checks the disk space before adding metrics to the disk queue.
# # check_disk_space = true
# ## Connection timeout, defaults to "5s" if not set.
# # timeout = "5s"
#
# ## The percent of used file system space at which the sender will stop queueing.
# ## When we will reach that percentage, the file system in which the queue is stored will drop
# ## all new logs until the percentage of used space drops below that threshold.
# # disk_threshold = 98
#
# ## How often Logz.io sender should drain the queue.
# ## Valid time units are "ns", "us" (or "µs"), "ms", "s", "m", "h".
# # drain_duration = "3s"
#
# ## Where Logz.io sender should store the queue
# ## queue_dir = Sprintf("%s%s%s%s%d", os.TempDir(), string(os.PathSeparator),
# ## "logzio-buffer", string(os.PathSeparator), time.Now().UnixNano())
# ## Optional TLS Config
# # tls_ca = "/etc/telegraf/ca.pem"
# # tls_cert = "/etc/telegraf/cert.pem"
# # tls_key = "/etc/telegraf/key.pem"
#
# ## Logz.io account token
# token = "your Logz.io token" # required
# token = "your logz.io token" # required
#
# ## Use your listener URL for your Logz.io account region.
# # url = "https://listener.logz.io:8071"
@ -3036,21 +3028,20 @@
# # Calculates a derivative for every field.
# [[aggregators.derivative]]
# ## Specific Derivative Aggregator Arguments:
#
# ## Configure a custom derivation variable. Timestamp is used if none is given.
# # variable = ""
#
# ## Suffix to add to the field name for the derivative name.
# # suffix = "_rate"
#
# ## Roll-Over last measurement to first measurement of next period
# # max_roll_over = 10
#
# ## General Aggregator Arguments:
#
# ## calculate derivative every 30 seconds
# ## The period in which to flush the aggregator.
# period = "30s"
# ##
# ## Suffix to append for the resulting derivative field.
# # suffix = "_rate"
# ##
# ## Field to use for the quotient when computing the derivative.
# ## When using a field as the derivation parameter the name of that field will
# ## be used for the resulting derivative, e.g. *fieldname_by_parameter*.
# ## By default the timestamps of the metrics are used and the suffix is omitted.
# # variable = ""
# ##
# ## Maximum number of roll-overs in case only one measurement is found during a period.
# # max_roll_over = 10
# # Report the final metric of a series
@ -6685,7 +6676,6 @@
# ## DEPRECATED: The 'snmp_legacy' plugin is deprecated in version 1.0.0, use 'inputs.snmp' instead.
# # DEPRECATED! PLEASE USE inputs.snmp INSTEAD.
# [[inputs.snmp_legacy]]
# ## Use 'oids.txt' file to translate oids to names
# ## To generate 'oids.txt' you need to run:
@ -7544,11 +7534,6 @@
# ## DEPRECATED: The 'cassandra' plugin is deprecated in version 1.7.0, use 'inputs.jolokia2' with the 'cassandra.conf' example configuration instead.
# # Read Cassandra metrics through Jolokia
# [[inputs.cassandra]]
# ## DEPRECATED: The cassandra plugin has been deprecated. Please use the
# ## jolokia2 plugin instead.
# ##
# ## see https://github.com/influxdata/telegraf/tree/master/plugins/inputs/jolokia2
#
# context = "/jolokia/read"
# ## List of cassandra servers exposing jolokia read service
# servers = ["myuser:mypassword@10.10.10.1:8778","10.10.10.2:8778",":8778"]
@ -8415,7 +8400,6 @@
# ## DEPRECATED: The 'kafka_consumer_legacy' plugin is deprecated in version 1.4.0, use 'inputs.kafka_consumer' instead, NOTE: 'kafka_consumer' only supports Kafka v0.8+.
# ## DEPRECATED: The 'kafka_consumer_legacy' plugin is deprecated in version 1.4.0, use 'inputs.kafka_consumer' instead, NOTE: 'kafka_consumer' only supports Kafka v0.8+.
# # Read metrics from Kafka topic(s)
# [[inputs.kafka_consumer_legacy]]
# ## topic(s) to consume
@ -8546,7 +8530,6 @@
# ## DEPRECATED: The 'logparser' plugin is deprecated in version 1.15.0, use 'inputs.tail' with 'grok' data format instead.
# ## DEPRECATED: The 'logparser' plugin is deprecated in version 1.15.0, use 'inputs.tail' with 'grok' data format instead.
# # Read metrics off Arista LANZ, via socket
# [[inputs.logparser]]
# ## Log files to parse.
@ -9027,7 +9010,7 @@
# # Riemann protobuff listener
# [[inputs.rimann_listener]]
# [[inputs.riemann_listener]]
# ## URL to listen on
# ## Default is "tcp://:5555"
# # service_address = "tcp://:8094"
@ -9620,7 +9603,201 @@
# # see https://github.com/influxdata/telegraf/tree/master/plugins/inputs/socket_listener
# vm_metric_exclude = [ "*" ]
# -# Read metrics from one or many vCenters
# [[inputs.vsphere]]
# ## List of vCenter URLs to be monitored. These three lines must be uncommented
# ## and edited for the plugin to work.
# vcenters = [ "https://vcenter.local/sdk" ]
# username = "user@corp.local"
# password = "secret"
#
# ## VMs
# ## Typical VM metrics (if omitted or empty, all metrics are collected)
# # vm_include = [ "/*/vm/**"] # Inventory path to VMs to collect (by default all are collected)
# # vm_exclude = [] # Inventory paths to exclude
# vm_metric_include = [
# "cpu.demand.average",
# "cpu.idle.summation",
# "cpu.latency.average",
# "cpu.readiness.average",
# "cpu.ready.summation",
# "cpu.run.summation",
# "cpu.usagemhz.average",
# "cpu.used.summation",
# "cpu.wait.summation",
# "mem.active.average",
# "mem.granted.average",
# "mem.latency.average",
# "mem.swapin.average",
# "mem.swapinRate.average",
# "mem.swapout.average",
# "mem.swapoutRate.average",
# "mem.usage.average",
# "mem.vmmemctl.average",
# "net.bytesRx.average",
# "net.bytesTx.average",
# "net.droppedRx.summation",
# "net.droppedTx.summation",
# "net.usage.average",
# "power.power.average",
# "virtualDisk.numberReadAveraged.average",
# "virtualDisk.numberWriteAveraged.average",
# "virtualDisk.read.average",
# "virtualDisk.readOIO.latest",
# "virtualDisk.throughput.usage.average",
# "virtualDisk.totalReadLatency.average",
# "virtualDisk.totalWriteLatency.average",
# "virtualDisk.write.average",
# "virtualDisk.writeOIO.latest",
# "sys.uptime.latest",
# ]
# # vm_metric_exclude = [] ## Nothing is excluded by default
# # vm_instances = true ## true by default
#
# ## Hosts
# ## Typical host metrics (if omitted or empty, all metrics are collected)
# # host_include = [ "/*/host/**"] # Inventory path to hosts to collect (by default all are collected)
# # host_exclude [] # Inventory paths to exclude
# host_metric_include = [
# "cpu.coreUtilization.average",
# "cpu.costop.summation",
# "cpu.demand.average",
# "cpu.idle.summation",
# "cpu.latency.average",
# "cpu.readiness.average",
# "cpu.ready.summation",
# "cpu.swapwait.summation",
# "cpu.usage.average",
# "cpu.usagemhz.average",
# "cpu.used.summation",
# "cpu.utilization.average",
# "cpu.wait.summation",
# "disk.deviceReadLatency.average",
# "disk.deviceWriteLatency.average",
# "disk.kernelReadLatency.average",
# "disk.kernelWriteLatency.average",
# "disk.numberReadAveraged.average",
# "disk.numberWriteAveraged.average",
# "disk.read.average",
# "disk.totalReadLatency.average",
# "disk.totalWriteLatency.average",
# "disk.write.average",
# "mem.active.average",
# "mem.latency.average",
# "mem.state.latest",
# "mem.swapin.average",
# "mem.swapinRate.average",
# "mem.swapout.average",
# "mem.swapoutRate.average",
# "mem.totalCapacity.average",
# "mem.usage.average",
# "mem.vmmemctl.average",
# "net.bytesRx.average",
# "net.bytesTx.average",
# "net.droppedRx.summation",
# "net.droppedTx.summation",
# "net.errorsRx.summation",
# "net.errorsTx.summation",
# "net.usage.average",
# "power.power.average",
# "storageAdapter.numberReadAveraged.average",
# "storageAdapter.numberWriteAveraged.average",
# "storageAdapter.read.average",
# "storageAdapter.write.average",
# "sys.uptime.latest",
# ]
# ## Collect IP addresses? Valid values are "ipv4" and "ipv6"
# # ip_addresses = ["ipv6", "ipv4" ]
#
# # host_metric_exclude = [] ## Nothing excluded by default
# # host_instances = true ## true by default
#
#
# ## Clusters
# # cluster_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
# # cluster_exclude = [] # Inventory paths to exclude
# # cluster_metric_include = [] ## if omitted or empty, all metrics are collected
# # cluster_metric_exclude = [] ## Nothing excluded by default
# # cluster_instances = false ## false by default
#
# ## Resource Pools
# # datastore_include = [ "/*/host/**"] # Inventory path to datastores to collect (by default all are collected)
# # datastore_exclude = [] # Inventory paths to exclude
# # datastore_metric_include = [] ## if omitted or empty, all metrics are collected
# # datastore_metric_exclude = [] ## Nothing excluded by default
# # datastore_instances = false ## false by default
#
# ## Datastores
# # datastore_include = [ "/*/datastore/**"] # Inventory path to datastores to collect (by default all are collected)
# # datastore_exclude = [] # Inventory paths to exclude
# # datastore_metric_include = [] ## if omitted or empty, all metrics are collected
# # datastore_metric_exclude = [] ## Nothing excluded by default
# # datastore_instances = false ## false by default
#
# ## Datacenters
# # datacenter_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
# # datacenter_exclude = [] # Inventory paths to exclude
# datacenter_metric_include = [] ## if omitted or empty, all metrics are collected
# datacenter_metric_exclude = [ "*" ] ## Datacenters are not collected by default.
# # datacenter_instances = false ## false by default
#
# ## Plugin Settings
# ## separator character to use for measurement and field names (default: "_")
# # separator = "_"
#
# ## number of objects to retrieve per query for realtime resources (vms and hosts)
# ## set to 64 for vCenter 5.5 and 6.0 (default: 256)
# # max_query_objects = 256
#
# ## number of metrics to retrieve per query for non-realtime resources (clusters and datastores)
# ## set to 64 for vCenter 5.5 and 6.0 (default: 256)
# # max_query_metrics = 256
#
# ## number of go routines to use for collection and discovery of objects and metrics
# # collect_concurrency = 1
# # discover_concurrency = 1
#
# ## the interval before (re)discovering objects subject to metrics collection (default: 300s)
# # object_discovery_interval = "300s"
#
# ## timeout applies to any of the api request made to vcenter
# # timeout = "60s"
#
# ## When set to true, all samples are sent as integers. This makes the output
# ## data types backwards compatible with Telegraf 1.9 or lower. Normally all
# ## samples from vCenter, with the exception of percentages, are integer
# ## values, but under some conditions, some averaging takes place internally in
# ## the plugin. Setting this flag to "false" will send values as floats to
# ## preserve the full precision when averaging takes place.
# # use_int_samples = true
#
# ## Custom attributes from vCenter can be very useful for queries in order to slice the
# ## metrics along different dimension and for forming ad-hoc relationships. They are disabled
# ## by default, since they can add a considerable amount of tags to the resulting metrics. To
# ## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
# ## to select the attributes you want to include.
# ## By default, since they can add a considerable amount of tags to the resulting metrics. To
# ## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
# ## to select the attributes you want to include.
# # custom_attribute_include = []
# # custom_attribute_exclude = ["*"]
#
# ## The number of vSphere 5 minute metric collection cycles to look back for non-realtime metrics. In
# ## some versions (6.7, 7.0 and possible more), certain metrics, such as cluster metrics, may be reported
# ## with a significant delay (>30min). If this happens, try increasing this number. Please note that increasing
# ## it too much may cause performance issues.
# # metric_lookback = 3
#
# ## Optional SSL Config
# # ssl_ca = "/path/to/cafile"
# # ssl_cert = "/path/to/certfile"
# # ssl_key = "/path/to/keyfile"
# ## Use SSL but skip chain & host verification
# # insecure_skip_verify = false
#
# ## The Historical Interval value must match EXACTLY the interval in the daily
# # "Interval Duration" found on the VCenter server under Configure > General > Statistics > Statistic intervals
# # historical_interval = "5m"
# # A Webhooks Event collector