chore(outputs): migrate sample configs into separate files (#11131)
This commit is contained in:
parent
4b3a5d5559
commit
256caede89
|
|
@ -0,0 +1,10 @@
|
||||||
|
# Configuration for Amon Server to send metrics to.
|
||||||
|
[[outputs.amon]]
|
||||||
|
## Amon Server Key
|
||||||
|
server_key = "my-server-key" # required.
|
||||||
|
|
||||||
|
## Amon Instance URL
|
||||||
|
amon_instance = "https://youramoninstance" # required
|
||||||
|
|
||||||
|
## Connection timeout.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
@ -0,0 +1,95 @@
|
||||||
|
# Publishes metrics to an AMQP broker
|
||||||
|
[[outputs.amqp]]
|
||||||
|
## Broker to publish to.
|
||||||
|
## deprecated in 1.7; use the brokers option
|
||||||
|
# url = "amqp://localhost:5672/influxdb"
|
||||||
|
|
||||||
|
## Brokers to publish to. If multiple brokers are specified a random broker
|
||||||
|
## will be selected anytime a connection is established. This can be
|
||||||
|
## helpful for load balancing when not using a dedicated load balancer.
|
||||||
|
brokers = ["amqp://localhost:5672/influxdb"]
|
||||||
|
|
||||||
|
## Maximum messages to send over a connection. Once this is reached, the
|
||||||
|
## connection is closed and a new connection is made. This can be helpful for
|
||||||
|
## load balancing when not using a dedicated load balancer.
|
||||||
|
# max_messages = 0
|
||||||
|
|
||||||
|
## Exchange to declare and publish to.
|
||||||
|
exchange = "telegraf"
|
||||||
|
|
||||||
|
## Exchange type; common types are "direct", "fanout", "topic", "header", "x-consistent-hash".
|
||||||
|
# exchange_type = "topic"
|
||||||
|
|
||||||
|
## If true, exchange will be passively declared.
|
||||||
|
# exchange_passive = false
|
||||||
|
|
||||||
|
## Exchange durability can be either "transient" or "durable".
|
||||||
|
# exchange_durability = "durable"
|
||||||
|
|
||||||
|
## Additional exchange arguments.
|
||||||
|
# exchange_arguments = { }
|
||||||
|
# exchange_arguments = {"hash_property" = "timestamp"}
|
||||||
|
|
||||||
|
## Authentication credentials for the PLAIN auth_method.
|
||||||
|
# username = ""
|
||||||
|
# password = ""
|
||||||
|
|
||||||
|
## Auth method. PLAIN and EXTERNAL are supported
|
||||||
|
## Using EXTERNAL requires enabling the rabbitmq_auth_mechanism_ssl plugin as
|
||||||
|
## described here: https://www.rabbitmq.com/plugins.html
|
||||||
|
# auth_method = "PLAIN"
|
||||||
|
|
||||||
|
## Metric tag to use as a routing key.
|
||||||
|
## ie, if this tag exists, its value will be used as the routing key
|
||||||
|
# routing_tag = "host"
|
||||||
|
|
||||||
|
## Static routing key. Used when no routing_tag is set or as a fallback
|
||||||
|
## when the tag specified in routing tag is not found.
|
||||||
|
# routing_key = ""
|
||||||
|
# routing_key = "telegraf"
|
||||||
|
|
||||||
|
## Delivery Mode controls if a published message is persistent.
|
||||||
|
## One of "transient" or "persistent".
|
||||||
|
# delivery_mode = "transient"
|
||||||
|
|
||||||
|
## InfluxDB database added as a message header.
|
||||||
|
## deprecated in 1.7; use the headers option
|
||||||
|
# database = "telegraf"
|
||||||
|
|
||||||
|
## InfluxDB retention policy added as a message header
|
||||||
|
## deprecated in 1.7; use the headers option
|
||||||
|
# retention_policy = "default"
|
||||||
|
|
||||||
|
## Static headers added to each published message.
|
||||||
|
# headers = { }
|
||||||
|
# headers = {"database" = "telegraf", "retention_policy" = "default"}
|
||||||
|
|
||||||
|
## Connection timeout. If not provided, will default to 5s. 0s means no
|
||||||
|
## timeout (not recommended).
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## If true use batch serialization format instead of line based delimiting.
|
||||||
|
## Only applies to data formats which are not line based such as JSON.
|
||||||
|
## Recommended to set to true.
|
||||||
|
# use_batch_format = false
|
||||||
|
|
||||||
|
## Content encoding for message payloads, can be set to "gzip" to or
|
||||||
|
## "identity" to apply no encoding.
|
||||||
|
##
|
||||||
|
## Please note that when use_batch_format = false each amqp message contains only
|
||||||
|
## a single metric, it is recommended to use compression with batch format
|
||||||
|
## for best results.
|
||||||
|
# content_encoding = "identity"
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
@ -0,0 +1,21 @@
|
||||||
|
# Send metrics to Azure Application Insights
|
||||||
|
[[outputs.application_insights]]
|
||||||
|
## Instrumentation key of the Application Insights resource.
|
||||||
|
instrumentation_key = "xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxx"
|
||||||
|
|
||||||
|
## Regions that require endpoint modification https://docs.microsoft.com/en-us/azure/azure-monitor/app/custom-endpoints
|
||||||
|
# endpoint_url = "https://dc.services.visualstudio.com/v2/track"
|
||||||
|
|
||||||
|
## Timeout for closing (default: 5s).
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Enable additional diagnostic logging.
|
||||||
|
# enable_diagnostic_logging = false
|
||||||
|
|
||||||
|
## Context Tag Sources add Application Insights context tags to a tag value.
|
||||||
|
##
|
||||||
|
## For list of allowed context tag keys see:
|
||||||
|
## https://github.com/microsoft/ApplicationInsights-Go/blob/master/appinsights/contracts/contexttagkeys.go
|
||||||
|
# [outputs.application_insights.context_tag_sources]
|
||||||
|
# "ai.cloud.role" = "kubernetes_container_name"
|
||||||
|
# "ai.cloud.roleInstance" = "kubernetes_pod_name"
|
||||||
|
|
@ -0,0 +1,25 @@
|
||||||
|
# Sends metrics to Azure Data Explorer
|
||||||
|
[[outputs.azure_data_explorer]]
|
||||||
|
## The URI property of the Azure Data Explorer resource on Azure
|
||||||
|
## ex: endpoint_url = https://myadxresource.australiasoutheast.kusto.windows.net
|
||||||
|
endpoint_url = ""
|
||||||
|
|
||||||
|
## The Azure Data Explorer database that the metrics will be ingested into.
|
||||||
|
## The plugin will NOT generate this database automatically, it's expected that this database already exists before ingestion.
|
||||||
|
## ex: "exampledatabase"
|
||||||
|
database = ""
|
||||||
|
|
||||||
|
## Timeout for Azure Data Explorer operations
|
||||||
|
# timeout = "20s"
|
||||||
|
|
||||||
|
## Type of metrics grouping used when pushing to Azure Data Explorer.
|
||||||
|
## Default is "TablePerMetric" for one table per different metric.
|
||||||
|
## For more information, please check the plugin README.
|
||||||
|
# metrics_grouping_type = "TablePerMetric"
|
||||||
|
|
||||||
|
## Name of the single table to store all the metrics (Only needed if metrics_grouping_type is "SingleTable").
|
||||||
|
# table_name = ""
|
||||||
|
|
||||||
|
## Creates tables and relevant mapping if set to true(default).
|
||||||
|
## Skips table and mapping creation if set to false, this is useful for running Telegraf with the lowest possible permissions i.e. table ingestor role.
|
||||||
|
# create_tables = true
|
||||||
|
|
@ -0,0 +1,29 @@
|
||||||
|
# Send aggregate metrics to Azure Monitor
|
||||||
|
[[outputs.azure_monitor]]
|
||||||
|
## Timeout for HTTP writes.
|
||||||
|
# timeout = "20s"
|
||||||
|
|
||||||
|
## Set the namespace prefix, defaults to "Telegraf/<input-name>".
|
||||||
|
# namespace_prefix = "Telegraf/"
|
||||||
|
|
||||||
|
## Azure Monitor doesn't have a string value type, so convert string
|
||||||
|
## fields to dimensions (a.k.a. tags) if enabled. Azure Monitor allows
|
||||||
|
## a maximum of 10 dimensions so Telegraf will only send the first 10
|
||||||
|
## alphanumeric dimensions.
|
||||||
|
# strings_as_dimensions = false
|
||||||
|
|
||||||
|
## Both region and resource_id must be set or be available via the
|
||||||
|
## Instance Metadata service on Azure Virtual Machines.
|
||||||
|
#
|
||||||
|
## Azure Region to publish metrics against.
|
||||||
|
## ex: region = "southcentralus"
|
||||||
|
# region = ""
|
||||||
|
#
|
||||||
|
## The Azure Resource ID against which metric will be logged, e.g.
|
||||||
|
## ex: resource_id = "/subscriptions/<subscription_id>/resourceGroups/<resource_group>/providers/Microsoft.Compute/virtualMachines/<vm_name>"
|
||||||
|
# resource_id = ""
|
||||||
|
|
||||||
|
## Optionally, if in Azure US Government, China, or other sovereign
|
||||||
|
## cloud environment, set the appropriate REST endpoint for receiving
|
||||||
|
## metrics. (Note: region may be unused in this context)
|
||||||
|
# endpoint_url = "https://monitoring.core.usgovcloudapi.net"
|
||||||
|
|
@ -0,0 +1,16 @@
|
||||||
|
# Configuration for Google Cloud BigQuery to send entries
|
||||||
|
[[outputs.bigquery]]
|
||||||
|
## Credentials File
|
||||||
|
credentials_file = "/path/to/service/account/key.json"
|
||||||
|
|
||||||
|
## Google Cloud Platform Project
|
||||||
|
project = "my-gcp-project"
|
||||||
|
|
||||||
|
## The namespace for the metric descriptor
|
||||||
|
dataset = "telegraf"
|
||||||
|
|
||||||
|
## Timeout for BigQuery operations.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Character to replace hyphens on Metric name
|
||||||
|
# replace_hyphen_to = "_"
|
||||||
|
|
@ -0,0 +1,49 @@
|
||||||
|
# Publish Telegraf metrics to a Google Cloud PubSub topic
|
||||||
|
[[outputs.cloud_pubsub]]
|
||||||
|
## Required. Name of Google Cloud Platform (GCP) Project that owns
|
||||||
|
## the given PubSub topic.
|
||||||
|
project = "my-project"
|
||||||
|
|
||||||
|
## Required. Name of PubSub topic to publish metrics to.
|
||||||
|
topic = "my-topic"
|
||||||
|
|
||||||
|
## Required. Data format to consume.
|
||||||
|
## Each data format has its own unique set of configuration options.
|
||||||
|
## Read more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
||||||
|
## Optional. Filepath for GCP credentials JSON file to authorize calls to
|
||||||
|
## PubSub APIs. If not set explicitly, Telegraf will attempt to use
|
||||||
|
## Application Default Credentials, which is preferred.
|
||||||
|
# credentials_file = "path/to/my/creds.json"
|
||||||
|
|
||||||
|
## Optional. If true, will send all metrics per write in one PubSub message.
|
||||||
|
# send_batched = true
|
||||||
|
|
||||||
|
## The following publish_* parameters specifically configures batching
|
||||||
|
## requests made to the GCP Cloud PubSub API via the PubSub Golang library. Read
|
||||||
|
## more here: https://godoc.org/cloud.google.com/go/pubsub#PublishSettings
|
||||||
|
|
||||||
|
## Optional. Send a request to PubSub (i.e. actually publish a batch)
|
||||||
|
## when it has this many PubSub messages. If send_batched is true,
|
||||||
|
## this is ignored and treated as if it were 1.
|
||||||
|
# publish_count_threshold = 1000
|
||||||
|
|
||||||
|
## Optional. Send a request to PubSub (i.e. actually publish a batch)
|
||||||
|
## when it has this many PubSub messages. If send_batched is true,
|
||||||
|
## this is ignored and treated as if it were 1
|
||||||
|
# publish_byte_threshold = 1000000
|
||||||
|
|
||||||
|
## Optional. Specifically configures requests made to the PubSub API.
|
||||||
|
# publish_num_go_routines = 2
|
||||||
|
|
||||||
|
## Optional. Specifies a timeout for requests to the PubSub API.
|
||||||
|
# publish_timeout = "30s"
|
||||||
|
|
||||||
|
## Optional. If true, published PubSub message data will be base64-encoded.
|
||||||
|
# base64_data = false
|
||||||
|
|
||||||
|
## Optional. PubSub attributes to add to metrics.
|
||||||
|
# [outputs.cloud_pubsub.attributes]
|
||||||
|
# my_attr = "tag_value"
|
||||||
|
|
@ -0,0 +1,42 @@
|
||||||
|
# Configuration for AWS CloudWatch output.
|
||||||
|
[[outputs.cloudwatch]]
|
||||||
|
## Amazon REGION
|
||||||
|
region = "us-east-1"
|
||||||
|
|
||||||
|
## Amazon Credentials
|
||||||
|
## Credentials are loaded in the following order
|
||||||
|
## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||||
|
## 2) Assumed credentials via STS if role_arn is specified
|
||||||
|
## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||||
|
## 4) shared profile from 'profile'
|
||||||
|
## 5) environment variables
|
||||||
|
## 6) shared credentials file
|
||||||
|
## 7) EC2 Instance Profile
|
||||||
|
#access_key = ""
|
||||||
|
#secret_key = ""
|
||||||
|
#token = ""
|
||||||
|
#role_arn = ""
|
||||||
|
#web_identity_token_file = ""
|
||||||
|
#role_session_name = ""
|
||||||
|
#profile = ""
|
||||||
|
#shared_credential_file = ""
|
||||||
|
|
||||||
|
## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
## determined and this option should only be set if you wish to override the
|
||||||
|
## default.
|
||||||
|
## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# endpoint_url = ""
|
||||||
|
|
||||||
|
## Namespace for the CloudWatch MetricDatums
|
||||||
|
namespace = "InfluxData/Telegraf"
|
||||||
|
|
||||||
|
## If you have a large amount of metrics, you should consider to send statistic
|
||||||
|
## values instead of raw metrics which could not only improve performance but
|
||||||
|
## also save AWS API cost. If enable this flag, this plugin would parse the required
|
||||||
|
## CloudWatch statistic fields (count, min, max, and sum) and send them to CloudWatch.
|
||||||
|
## You could use basicstats aggregator to calculate those fields. If not all statistic
|
||||||
|
## fields are available, all fields would still be sent as raw metrics.
|
||||||
|
# write_statistics = false
|
||||||
|
|
||||||
|
## Enable high resolution metrics of 1 second (if not enabled, standard resolution are of 60 seconds precision)
|
||||||
|
# high_resolution_metrics = false
|
||||||
|
|
@ -0,0 +1,58 @@
|
||||||
|
# Configuration for AWS CloudWatchLogs output.
|
||||||
|
[[outputs.cloudwatch_logs]]
|
||||||
|
## The region is the Amazon region that you wish to connect to.
|
||||||
|
## Examples include but are not limited to:
|
||||||
|
## - us-west-1
|
||||||
|
## - us-west-2
|
||||||
|
## - us-east-1
|
||||||
|
## - ap-southeast-1
|
||||||
|
## - ap-southeast-2
|
||||||
|
## ...
|
||||||
|
region = "us-east-1"
|
||||||
|
|
||||||
|
## Amazon Credentials
|
||||||
|
## Credentials are loaded in the following order
|
||||||
|
## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||||
|
## 2) Assumed credentials via STS if role_arn is specified
|
||||||
|
## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||||
|
## 4) shared profile from 'profile'
|
||||||
|
## 5) environment variables
|
||||||
|
## 6) shared credentials file
|
||||||
|
## 7) EC2 Instance Profile
|
||||||
|
#access_key = ""
|
||||||
|
#secret_key = ""
|
||||||
|
#token = ""
|
||||||
|
#role_arn = ""
|
||||||
|
#web_identity_token_file = ""
|
||||||
|
#role_session_name = ""
|
||||||
|
#profile = ""
|
||||||
|
#shared_credential_file = ""
|
||||||
|
|
||||||
|
## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
## determined and this option should only be set if you wish to override the
|
||||||
|
## default.
|
||||||
|
## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# endpoint_url = ""
|
||||||
|
|
||||||
|
## Cloud watch log group. Must be created in AWS cloudwatch logs upfront!
|
||||||
|
## For example, you can specify the name of the k8s cluster here to group logs from all cluster in oine place
|
||||||
|
log_group = "my-group-name"
|
||||||
|
|
||||||
|
## Log stream in log group
|
||||||
|
## Either log group name or reference to metric attribute, from which it can be parsed:
|
||||||
|
## tag:<TAG_NAME> or field:<FIELD_NAME>. If log stream is not exist, it will be created.
|
||||||
|
## Since AWS is not automatically delete logs streams with expired logs entries (i.e. empty log stream)
|
||||||
|
## you need to put in place appropriate house-keeping (https://forums.aws.amazon.com/thread.jspa?threadID=178855)
|
||||||
|
log_stream = "tag:location"
|
||||||
|
|
||||||
|
## Source of log data - metric name
|
||||||
|
## specify the name of the metric, from which the log data should be retrieved.
|
||||||
|
## I.e., if you are using docker_log plugin to stream logs from container, then
|
||||||
|
## specify log_data_metric_name = "docker_log"
|
||||||
|
log_data_metric_name = "docker_log"
|
||||||
|
|
||||||
|
## Specify from which metric attribute the log data should be retrieved:
|
||||||
|
## tag:<TAG_NAME> or field:<FIELD_NAME>.
|
||||||
|
## I.e., if you are using docker_log plugin to stream logs from container, then
|
||||||
|
## specify log_data_source = "field:message"
|
||||||
|
log_data_source = "field:message"
|
||||||
|
|
@ -0,0 +1,13 @@
|
||||||
|
# Configuration for CrateDB to send metrics to.
|
||||||
|
[[outputs.cratedb]]
|
||||||
|
# A github.com/jackc/pgx/v4 connection string.
|
||||||
|
# See https://pkg.go.dev/github.com/jackc/pgx/v4#ParseConfig
|
||||||
|
url = "postgres://user:password@localhost/schema?sslmode=disable"
|
||||||
|
# Timeout for all CrateDB queries.
|
||||||
|
timeout = "5s"
|
||||||
|
# Name of the table to store metrics in.
|
||||||
|
table = "metrics"
|
||||||
|
# If true, and the metrics table does not exist, create it automatically.
|
||||||
|
table_create = true
|
||||||
|
# The character(s) to replace any '.' in an object key with
|
||||||
|
key_separator = "_"
|
||||||
|
|
@ -0,0 +1,17 @@
|
||||||
|
# Configuration for DataDog API to send metrics to.
|
||||||
|
[[outputs.datadog]]
|
||||||
|
## Datadog API key
|
||||||
|
apikey = "my-secret-key"
|
||||||
|
|
||||||
|
## Connection timeout.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Write URL override; useful for debugging.
|
||||||
|
# url = "https://app.datadoghq.com/api/v1/series"
|
||||||
|
|
||||||
|
## Set http_proxy (telegraf uses the system wide proxy settings if it isn't set)
|
||||||
|
# http_proxy_url = "http://localhost:8888"
|
||||||
|
|
||||||
|
## Override the default (none) compression used to send data.
|
||||||
|
## Supports: "zlib", "none"
|
||||||
|
# compression = "none"
|
||||||
|
|
@ -0,0 +1,3 @@
|
||||||
|
# Send metrics to nowhere at all
|
||||||
|
[[outputs.discard]]
|
||||||
|
# no configuration
|
||||||
|
|
@ -0,0 +1,36 @@
|
||||||
|
# Send telegraf metrics to a Dynatrace environment
|
||||||
|
[[outputs.dynatrace]]
|
||||||
|
## For usage with the Dynatrace OneAgent you can omit any configuration,
|
||||||
|
## the only requirement is that the OneAgent is running on the same host.
|
||||||
|
## Only setup environment url and token if you want to monitor a Host without the OneAgent present.
|
||||||
|
##
|
||||||
|
## Your Dynatrace environment URL.
|
||||||
|
## For Dynatrace OneAgent you can leave this empty or set it to "http://127.0.0.1:14499/metrics/ingest" (default)
|
||||||
|
## For Dynatrace SaaS environments the URL scheme is "https://{your-environment-id}.live.dynatrace.com/api/v2/metrics/ingest"
|
||||||
|
## For Dynatrace Managed environments the URL scheme is "https://{your-domain}/e/{your-environment-id}/api/v2/metrics/ingest"
|
||||||
|
url = ""
|
||||||
|
|
||||||
|
## Your Dynatrace API token.
|
||||||
|
## Create an API token within your Dynatrace environment, by navigating to Settings > Integration > Dynatrace API
|
||||||
|
## The API token needs data ingest scope permission. When using OneAgent, no API token is required.
|
||||||
|
api_token = ""
|
||||||
|
|
||||||
|
## Optional prefix for metric names (e.g.: "telegraf")
|
||||||
|
prefix = "telegraf"
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Optional flag for ignoring tls certificate check
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Connection timeout, defaults to "5s" if not set.
|
||||||
|
timeout = "5s"
|
||||||
|
|
||||||
|
## If you want metrics to be treated and reported as delta counters, add the metric names here
|
||||||
|
additional_counters = [ ]
|
||||||
|
|
||||||
|
## Optional dimensions to be added to every metric
|
||||||
|
# [outputs.dynatrace.default_dimensions]
|
||||||
|
# default_key = "default value"
|
||||||
|
|
@ -0,0 +1,79 @@
|
||||||
|
# Configuration for Elasticsearch to send metrics to.
|
||||||
|
[[outputs.elasticsearch]]
|
||||||
|
## The full HTTP endpoint URL for your Elasticsearch instance
|
||||||
|
## Multiple urls can be specified as part of the same cluster,
|
||||||
|
## this means that only ONE of the urls will be written to each interval
|
||||||
|
urls = [ "http://node1.es.example.com:9200" ] # required.
|
||||||
|
## Elasticsearch client timeout, defaults to "5s" if not set.
|
||||||
|
timeout = "5s"
|
||||||
|
## Set to true to ask Elasticsearch a list of all cluster nodes,
|
||||||
|
## thus it is not necessary to list all nodes in the urls config option
|
||||||
|
enable_sniffer = false
|
||||||
|
## Set to true to enable gzip compression
|
||||||
|
enable_gzip = false
|
||||||
|
## Set the interval to check if the Elasticsearch nodes are available
|
||||||
|
## Setting to "0s" will disable the health check (not recommended in production)
|
||||||
|
health_check_interval = "10s"
|
||||||
|
## Set the timeout for periodic health checks.
|
||||||
|
# health_check_timeout = "1s"
|
||||||
|
## HTTP basic authentication details.
|
||||||
|
## HTTP basic authentication details
|
||||||
|
# username = "telegraf"
|
||||||
|
# password = "mypassword"
|
||||||
|
## HTTP bearer token authentication details
|
||||||
|
# auth_bearer_token = "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9"
|
||||||
|
|
||||||
|
## Index Config
|
||||||
|
## The target index for metrics (Elasticsearch will create if it not exists).
|
||||||
|
## You can use the date specifiers below to create indexes per time frame.
|
||||||
|
## The metric timestamp will be used to decide the destination index name
|
||||||
|
# %Y - year (2016)
|
||||||
|
# %y - last two digits of year (00..99)
|
||||||
|
# %m - month (01..12)
|
||||||
|
# %d - day of month (e.g., 01)
|
||||||
|
# %H - hour (00..23)
|
||||||
|
# %V - week of the year (ISO week) (01..53)
|
||||||
|
## Additionally, you can specify a tag name using the notation {{tag_name}}
|
||||||
|
## which will be used as part of the index name. If the tag does not exist,
|
||||||
|
## the default tag value will be used.
|
||||||
|
# index_name = "telegraf-{{host}}-%Y.%m.%d"
|
||||||
|
# default_tag_value = "none"
|
||||||
|
index_name = "telegraf-%Y.%m.%d" # required.
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Template Config
|
||||||
|
## Set to true if you want telegraf to manage its index template.
|
||||||
|
## If enabled it will create a recommended index template for telegraf indexes
|
||||||
|
manage_template = true
|
||||||
|
## The template name used for telegraf indexes
|
||||||
|
template_name = "telegraf"
|
||||||
|
## Set to true if you want telegraf to overwrite an existing template
|
||||||
|
overwrite_template = false
|
||||||
|
## If set to true a unique ID hash will be sent as sha256(concat(timestamp,measurement,series-hash)) string
|
||||||
|
## it will enable data resend and update metric points avoiding duplicated metrics with diferent id's
|
||||||
|
force_document_id = false
|
||||||
|
|
||||||
|
## Specifies the handling of NaN and Inf values.
|
||||||
|
## This option can have the following values:
|
||||||
|
## none -- do not modify field-values (default); will produce an error if NaNs or infs are encountered
|
||||||
|
## drop -- drop fields containing NaNs or infs
|
||||||
|
## replace -- replace with the value in "float_replacement_value" (default: 0.0)
|
||||||
|
## NaNs and inf will be replaced with the given number, -inf with the negative of that number
|
||||||
|
# float_handling = "none"
|
||||||
|
# float_replacement_value = 0.0
|
||||||
|
|
||||||
|
## Pipeline Config
|
||||||
|
## To use a ingest pipeline, set this to the name of the pipeline you want to use.
|
||||||
|
# use_pipeline = "my_pipeline"
|
||||||
|
## Additionally, you can specify a tag name using the notation {{tag_name}}
|
||||||
|
## which will be used as part of the pipeline name. If the tag does not exist,
|
||||||
|
## the default pipeline will be used as the pipeline. If no default pipeline is set,
|
||||||
|
## no pipeline is used for the metric.
|
||||||
|
# use_pipeline = "{{es_pipeline}}"
|
||||||
|
# default_pipeline = "my_pipeline"
|
||||||
|
|
@ -0,0 +1,12 @@
|
||||||
|
# Configuration for Event Hubs output plugin
|
||||||
|
[[outputs.event_hubs]]
|
||||||
|
## The full connection string to the Event Hub (required)
|
||||||
|
## The shared access key must have "Send" permissions on the target Event Hub.
|
||||||
|
connection_string = "Endpoint=sb://namespace.servicebus.windows.net/;SharedAccessKeyName=RootManageSharedAccessKey;SharedAccessKey=superSecret1234=;EntityPath=hubName"
|
||||||
|
## Client timeout (defaults to 30s)
|
||||||
|
# timeout = "30s"
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "json"
|
||||||
|
|
@ -0,0 +1,19 @@
|
||||||
|
# Send metrics to command as input over stdin
|
||||||
|
[[outputs.exec]]
|
||||||
|
## Command to ingest metrics via stdin.
|
||||||
|
command = ["tee", "-a", "/dev/null"]
|
||||||
|
|
||||||
|
## Environment variables
|
||||||
|
## Array of "key=value" pairs to pass as environment variables
|
||||||
|
## e.g. "KEY=value", "USERNAME=John Doe",
|
||||||
|
## "LD_LIBRARY_PATH=/opt/custom/lib64:/usr/local/libs"
|
||||||
|
# environment = []
|
||||||
|
|
||||||
|
## Timeout for command to complete.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
@ -0,0 +1,20 @@
|
||||||
|
# Run executable as long-running output plugin
|
||||||
|
[[outputs.execd]]
|
||||||
|
## One program to run as daemon.
|
||||||
|
## NOTE: process and each argument should each be their own string
|
||||||
|
command = ["my-telegraf-output", "--some-flag", "value"]
|
||||||
|
|
||||||
|
## Environment variables
|
||||||
|
## Array of "key=value" pairs to pass as environment variables
|
||||||
|
## e.g. "KEY=value", "USERNAME=John Doe",
|
||||||
|
## "LD_LIBRARY_PATH=/opt/custom/lib64:/usr/local/libs"
|
||||||
|
# environment = []
|
||||||
|
|
||||||
|
## Delay before the process is restarted after an unexpected termination
|
||||||
|
restart_delay = "10s"
|
||||||
|
|
||||||
|
## Data format to export.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
@ -0,0 +1,27 @@
|
||||||
|
# Send telegraf metrics to file(s)
|
||||||
|
[[outputs.file]]
|
||||||
|
## Files to write to, "stdout" is a specially handled file.
|
||||||
|
files = ["stdout", "/tmp/metrics.out"]
|
||||||
|
|
||||||
|
## Use batch serialization format instead of line based delimiting. The
|
||||||
|
## batch format allows for the production of non line based output formats and
|
||||||
|
## may more efficiently encode and write metrics.
|
||||||
|
# use_batch_format = false
|
||||||
|
|
||||||
|
## The file will be rotated after the time interval specified. When set
|
||||||
|
## to 0 no time based rotation is performed.
|
||||||
|
# rotation_interval = "0h"
|
||||||
|
|
||||||
|
## The logfile will be rotated when it becomes larger than the specified
|
||||||
|
## size. When set to 0 no size based rotation is performed.
|
||||||
|
# rotation_max_size = "0MB"
|
||||||
|
|
||||||
|
## Maximum number of rotated archives to keep, any older logs are deleted.
|
||||||
|
## If set to -1, no archives are removed.
|
||||||
|
# rotation_max_archives = 5
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
@ -0,0 +1,42 @@
|
||||||
|
# Configuration for Graphite server to send metrics to
|
||||||
|
[[outputs.graphite]]
|
||||||
|
## TCP endpoint for your graphite instance.
|
||||||
|
## If multiple endpoints are configured, the output will be load balanced.
|
||||||
|
## Only one of the endpoints will be written to with each iteration.
|
||||||
|
servers = ["localhost:2003"]
|
||||||
|
## Prefix metrics name
|
||||||
|
prefix = ""
|
||||||
|
## Graphite output template
|
||||||
|
## see https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
template = "host.tags.measurement.field"
|
||||||
|
|
||||||
|
## Enable Graphite tags support
|
||||||
|
# graphite_tag_support = false
|
||||||
|
|
||||||
|
## Define how metric names and tags are sanitized; options are "strict", or "compatible"
|
||||||
|
## strict - Default method, and backwards compatible with previous versionf of Telegraf
|
||||||
|
## compatible - More relaxed sanitizing when using tags, and compatible with the graphite spec
|
||||||
|
# graphite_tag_sanitize_mode = "strict"
|
||||||
|
|
||||||
|
## Character for separating metric name and field for Graphite tags
|
||||||
|
# graphite_separator = "."
|
||||||
|
|
||||||
|
## Graphite templates patterns
|
||||||
|
## 1. Template for cpu
|
||||||
|
## 2. Template for disk*
|
||||||
|
## 3. Default template
|
||||||
|
# templates = [
|
||||||
|
# "cpu tags.measurement.host.field",
|
||||||
|
# "disk* measurement.field",
|
||||||
|
# "host.measurement.tags.field"
|
||||||
|
#]
|
||||||
|
|
||||||
|
## timeout in seconds for the write connection to graphite
|
||||||
|
timeout = 2
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
@ -0,0 +1,24 @@
|
||||||
|
# Send telegraf metrics to graylog
|
||||||
|
[[outputs.graylog]]
|
||||||
|
## Endpoints for your graylog instances.
|
||||||
|
servers = ["udp://127.0.0.1:12201"]
|
||||||
|
|
||||||
|
## Connection timeout.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## The field to use as the GELF short_message, if unset the static string
|
||||||
|
## "telegraf" will be used.
|
||||||
|
## example: short_message_field = "message"
|
||||||
|
# short_message_field = ""
|
||||||
|
|
||||||
|
## According to GELF payload specification, additional fields names must be prefixed
|
||||||
|
## with an underscore. Previous versions did not prefix custom field 'name' with underscore.
|
||||||
|
## Set to true for backward compatibility.
|
||||||
|
# name_field_no_prefix = false
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
@ -0,0 +1,23 @@
|
||||||
|
# Send telegraf metrics to GroundWork Monitor
|
||||||
|
[[outputs.groundwork]]
|
||||||
|
## URL of your groundwork instance.
|
||||||
|
url = "https://groundwork.example.com"
|
||||||
|
|
||||||
|
## Agent uuid for GroundWork API Server.
|
||||||
|
agent_id = ""
|
||||||
|
|
||||||
|
## Username and password to access GroundWork API.
|
||||||
|
username = ""
|
||||||
|
password = ""
|
||||||
|
|
||||||
|
## Default display name for the host with services(metrics).
|
||||||
|
# default_host = "telegraf"
|
||||||
|
|
||||||
|
## Default service state.
|
||||||
|
# default_service_state = "SERVICE_OK"
|
||||||
|
|
||||||
|
## The name of the tag that contains the hostname.
|
||||||
|
# resource_tag = "host"
|
||||||
|
|
||||||
|
## The name of the tag that contains the host group name.
|
||||||
|
# group_tag = "group"
|
||||||
|
|
@ -0,0 +1,38 @@
|
||||||
|
# Configurable HTTP health check resource based on metrics
|
||||||
|
[[outputs.health]]
|
||||||
|
## Address and port to listen on.
|
||||||
|
## ex: service_address = "http://localhost:8080"
|
||||||
|
## service_address = "unix:///var/run/telegraf-health.sock"
|
||||||
|
# service_address = "http://:8080"
|
||||||
|
|
||||||
|
## The maximum duration for reading the entire request.
|
||||||
|
# read_timeout = "5s"
|
||||||
|
## The maximum duration for writing the entire response.
|
||||||
|
# write_timeout = "5s"
|
||||||
|
|
||||||
|
## Username and password to accept for HTTP basic authentication.
|
||||||
|
# basic_username = "user1"
|
||||||
|
# basic_password = "secret"
|
||||||
|
|
||||||
|
## Allowed CA certificates for client certificates.
|
||||||
|
# tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]
|
||||||
|
|
||||||
|
## TLS server certificate and private key.
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
|
||||||
|
## One or more check sub-tables should be defined, it is also recommended to
|
||||||
|
## use metric filtering to limit the metrics that flow into this output.
|
||||||
|
##
|
||||||
|
## When using the default buffer sizes, this example will fail when the
|
||||||
|
## metric buffer is half full.
|
||||||
|
##
|
||||||
|
## namepass = ["internal_write"]
|
||||||
|
## tagpass = { output = ["influxdb"] }
|
||||||
|
##
|
||||||
|
## [[outputs.health.compares]]
|
||||||
|
## field = "buffer_size"
|
||||||
|
## lt = 5000.0
|
||||||
|
##
|
||||||
|
## [[outputs.health.contains]]
|
||||||
|
## field = "buffer_size"
|
||||||
|
|
@ -0,0 +1,95 @@
|
||||||
|
# A plugin that can transmit metrics over HTTP
|
||||||
|
[[outputs.http]]
|
||||||
|
## URL is the address to send metrics to
|
||||||
|
url = "http://127.0.0.1:8080/telegraf"
|
||||||
|
|
||||||
|
## Timeout for HTTP message
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## HTTP method, one of: "POST" or "PUT"
|
||||||
|
# method = "POST"
|
||||||
|
|
||||||
|
## HTTP Basic Auth credentials
|
||||||
|
# username = "username"
|
||||||
|
# password = "pa$$word"
|
||||||
|
|
||||||
|
## OAuth2 Client Credentials Grant
|
||||||
|
# client_id = "clientid"
|
||||||
|
# client_secret = "secret"
|
||||||
|
# token_url = "https://indentityprovider/oauth2/v1/token"
|
||||||
|
# scopes = ["urn:opc:idm:__myscopes__"]
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Optional Cookie authentication
|
||||||
|
# cookie_auth_url = "https://localhost/authMe"
|
||||||
|
# cookie_auth_method = "POST"
|
||||||
|
# cookie_auth_username = "username"
|
||||||
|
# cookie_auth_password = "pa$$word"
|
||||||
|
# cookie_auth_headers = '{"Content-Type": "application/json", "X-MY-HEADER":"hello"}'
|
||||||
|
# cookie_auth_body = '{"username": "user", "password": "pa$$word", "authenticate": "me"}'
|
||||||
|
## cookie_auth_renewal not set or set to "0" will auth once and never renew the cookie
|
||||||
|
# cookie_auth_renewal = "5m"
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has it's own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
||||||
|
## Use batch serialization format (default) instead of line based format.
|
||||||
|
## Batch format is more efficient and should be used unless line based
|
||||||
|
## format is really needed.
|
||||||
|
# use_batch_format = true
|
||||||
|
|
||||||
|
## HTTP Content-Encoding for write request body, can be set to "gzip" to
|
||||||
|
## compress body or "identity" to apply no encoding.
|
||||||
|
# content_encoding = "identity"
|
||||||
|
|
||||||
|
## Additional HTTP headers
|
||||||
|
# [outputs.http.headers]
|
||||||
|
# # Should be set manually to "application/json" for json data_format
|
||||||
|
# Content-Type = "text/plain; charset=utf-8"
|
||||||
|
|
||||||
|
## MaxIdleConns controls the maximum number of idle (keep-alive)
|
||||||
|
## connections across all hosts. Zero means no limit.
|
||||||
|
# max_idle_conn = 0
|
||||||
|
|
||||||
|
## MaxIdleConnsPerHost, if non-zero, controls the maximum idle
|
||||||
|
## (keep-alive) connections to keep per-host. If zero,
|
||||||
|
## DefaultMaxIdleConnsPerHost is used(2).
|
||||||
|
# max_idle_conn_per_host = 2
|
||||||
|
|
||||||
|
## Idle (keep-alive) connection timeout.
|
||||||
|
## Maximum amount of time before idle connection is closed.
|
||||||
|
## Zero means no limit.
|
||||||
|
# idle_conn_timeout = 0
|
||||||
|
|
||||||
|
## Amazon Region
|
||||||
|
#region = "us-east-1"
|
||||||
|
|
||||||
|
## Amazon Credentials
|
||||||
|
## Credentials are loaded in the following order
|
||||||
|
## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||||
|
## 2) Assumed credentials via STS if role_arn is specified
|
||||||
|
## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||||
|
## 4) shared profile from 'profile'
|
||||||
|
## 5) environment variables
|
||||||
|
## 6) shared credentials file
|
||||||
|
## 7) EC2 Instance Profile
|
||||||
|
#access_key = ""
|
||||||
|
#secret_key = ""
|
||||||
|
#token = ""
|
||||||
|
#role_arn = ""
|
||||||
|
#web_identity_token_file = ""
|
||||||
|
#role_session_name = ""
|
||||||
|
#profile = ""
|
||||||
|
#shared_credential_file = ""
|
||||||
|
|
||||||
|
## Optional list of statuscodes (<200 or >300) upon which requests should not be retried
|
||||||
|
# non_retryable_statuscodes = [409, 413]
|
||||||
|
|
@ -0,0 +1,77 @@
|
||||||
|
# Configuration for sending metrics to InfluxDB
|
||||||
|
[[outputs.influxdb]]
|
||||||
|
## The full HTTP or UDP URL for your InfluxDB instance.
|
||||||
|
##
|
||||||
|
## Multiple URLs can be specified for a single cluster, only ONE of the
|
||||||
|
## urls will be written to each interval.
|
||||||
|
# urls = ["unix:///var/run/influxdb.sock"]
|
||||||
|
# urls = ["udp://127.0.0.1:8089"]
|
||||||
|
# urls = ["http://127.0.0.1:8086"]
|
||||||
|
|
||||||
|
## The target database for metrics; will be created as needed.
|
||||||
|
## For UDP url endpoint database needs to be configured on server side.
|
||||||
|
# database = "telegraf"
|
||||||
|
|
||||||
|
## The value of this tag will be used to determine the database. If this
|
||||||
|
## tag is not set the 'database' option is used as the default.
|
||||||
|
# database_tag = ""
|
||||||
|
|
||||||
|
## If true, the 'database_tag' will not be included in the written metric.
|
||||||
|
# exclude_database_tag = false
|
||||||
|
|
||||||
|
## If true, no CREATE DATABASE queries will be sent. Set to true when using
|
||||||
|
## Telegraf with a user without permissions to create databases or when the
|
||||||
|
## database already exists.
|
||||||
|
# skip_database_creation = false
|
||||||
|
|
||||||
|
## Name of existing retention policy to write to. Empty string writes to
|
||||||
|
## the default retention policy. Only takes effect when using HTTP.
|
||||||
|
# retention_policy = ""
|
||||||
|
|
||||||
|
## The value of this tag will be used to determine the retention policy. If this
|
||||||
|
## tag is not set the 'retention_policy' option is used as the default.
|
||||||
|
# retention_policy_tag = ""
|
||||||
|
|
||||||
|
## If true, the 'retention_policy_tag' will not be included in the written metric.
|
||||||
|
# exclude_retention_policy_tag = false
|
||||||
|
|
||||||
|
## Write consistency (clusters only), can be: "any", "one", "quorum", "all".
|
||||||
|
## Only takes effect when using HTTP.
|
||||||
|
# write_consistency = "any"
|
||||||
|
|
||||||
|
## Timeout for HTTP messages.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## HTTP Basic Auth
|
||||||
|
# username = "telegraf"
|
||||||
|
# password = "metricsmetricsmetricsmetrics"
|
||||||
|
|
||||||
|
## HTTP User-Agent
|
||||||
|
# user_agent = "telegraf"
|
||||||
|
|
||||||
|
## UDP payload size is the maximum packet size to send.
|
||||||
|
# udp_payload = "512B"
|
||||||
|
|
||||||
|
## Optional TLS Config for use on HTTP connections.
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## HTTP Proxy override, if unset values the standard proxy environment
|
||||||
|
## variables are consulted to determine which proxy, if any, should be used.
|
||||||
|
# http_proxy = "http://corporate.proxy:3128"
|
||||||
|
|
||||||
|
## Additional HTTP headers
|
||||||
|
# http_headers = {"X-Special-Header" = "Special-Value"}
|
||||||
|
|
||||||
|
## HTTP Content-Encoding for write request body, can be set to "gzip" to
|
||||||
|
## compress body or "identity" to apply no encoding.
|
||||||
|
# content_encoding = "gzip"
|
||||||
|
|
||||||
|
## When true, Telegraf will output unsigned integers as unsigned values,
|
||||||
|
## i.e.: "42u". You will need a version of InfluxDB supporting unsigned
|
||||||
|
## integer values. Enabling this option will result in field type errors if
|
||||||
|
## existing data has been written.
|
||||||
|
# influx_uint_support = false
|
||||||
|
|
@ -0,0 +1,51 @@
|
||||||
|
# Configuration for sending metrics to InfluxDB 2.0
|
||||||
|
[[outputs.influxdb_v2]]
|
||||||
|
## The URLs of the InfluxDB cluster nodes.
|
||||||
|
##
|
||||||
|
## Multiple URLs can be specified for a single cluster, only ONE of the
|
||||||
|
## urls will be written to each interval.
|
||||||
|
## ex: urls = ["https://us-west-2-1.aws.cloud2.influxdata.com"]
|
||||||
|
urls = ["http://127.0.0.1:8086"]
|
||||||
|
|
||||||
|
## Token for authentication.
|
||||||
|
token = ""
|
||||||
|
|
||||||
|
## Organization is the name of the organization you wish to write to.
|
||||||
|
organization = ""
|
||||||
|
|
||||||
|
## Destination bucket to write into.
|
||||||
|
bucket = ""
|
||||||
|
|
||||||
|
## The value of this tag will be used to determine the bucket. If this
|
||||||
|
## tag is not set the 'bucket' option is used as the default.
|
||||||
|
# bucket_tag = ""
|
||||||
|
|
||||||
|
## If true, the bucket tag will not be added to the metric.
|
||||||
|
# exclude_bucket_tag = false
|
||||||
|
|
||||||
|
## Timeout for HTTP messages.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Additional HTTP headers
|
||||||
|
# http_headers = {"X-Special-Header" = "Special-Value"}
|
||||||
|
|
||||||
|
## HTTP Proxy override, if unset values the standard proxy environment
|
||||||
|
## variables are consulted to determine which proxy, if any, should be used.
|
||||||
|
# http_proxy = "http://corporate.proxy:3128"
|
||||||
|
|
||||||
|
## HTTP User-Agent
|
||||||
|
# user_agent = "telegraf"
|
||||||
|
|
||||||
|
## Content-Encoding for write request body, can be set to "gzip" to
|
||||||
|
## compress body or "identity" to apply no encoding.
|
||||||
|
# content_encoding = "gzip"
|
||||||
|
|
||||||
|
## Enable or disable uint support for writing uints influxdb 2.0.
|
||||||
|
# influx_uint_support = false
|
||||||
|
|
||||||
|
## Optional TLS Config for use on HTTP connections.
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
@ -0,0 +1,13 @@
|
||||||
|
# Configuration for sending metrics to an Instrumental project
|
||||||
|
[[outputs.instrumental]]
|
||||||
|
## Project API Token (required)
|
||||||
|
api_token = "API Token" # required
|
||||||
|
## Prefix the metrics with a given name
|
||||||
|
prefix = ""
|
||||||
|
## Stats output template (Graphite formatting)
|
||||||
|
## see https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md#graphite
|
||||||
|
template = "host.tags.measurement.field"
|
||||||
|
## Timeout in seconds to connect
|
||||||
|
timeout = "2s"
|
||||||
|
## Debug true - Print communication to Instrumental
|
||||||
|
debug = false
|
||||||
|
|
@ -0,0 +1,151 @@
|
||||||
|
# Configuration for the Kafka server to send metrics to
|
||||||
|
[[outputs.kafka]]
|
||||||
|
## URLs of kafka brokers
|
||||||
|
brokers = ["localhost:9092"]
|
||||||
|
## Kafka topic for producer messages
|
||||||
|
topic = "telegraf"
|
||||||
|
|
||||||
|
## The value of this tag will be used as the topic. If not set the 'topic'
|
||||||
|
## option is used.
|
||||||
|
# topic_tag = ""
|
||||||
|
|
||||||
|
## If true, the 'topic_tag' will be removed from to the metric.
|
||||||
|
# exclude_topic_tag = false
|
||||||
|
|
||||||
|
## Optional Client id
|
||||||
|
# client_id = "Telegraf"
|
||||||
|
|
||||||
|
## Set the minimal supported Kafka version. Setting this enables the use of new
|
||||||
|
## Kafka features and APIs. Of particular interested, lz4 compression
|
||||||
|
## requires at least version 0.10.0.0.
|
||||||
|
## ex: version = "1.1.0"
|
||||||
|
# version = ""
|
||||||
|
|
||||||
|
## Optional topic suffix configuration.
|
||||||
|
## If the section is omitted, no suffix is used.
|
||||||
|
## Following topic suffix methods are supported:
|
||||||
|
## measurement - suffix equals to separator + measurement's name
|
||||||
|
## tags - suffix equals to separator + specified tags' values
|
||||||
|
## interleaved with separator
|
||||||
|
|
||||||
|
## Suffix equals to "_" + measurement name
|
||||||
|
# [outputs.kafka.topic_suffix]
|
||||||
|
# method = "measurement"
|
||||||
|
# separator = "_"
|
||||||
|
|
||||||
|
## Suffix equals to "__" + measurement's "foo" tag value.
|
||||||
|
## If there's no such a tag, suffix equals to an empty string
|
||||||
|
# [outputs.kafka.topic_suffix]
|
||||||
|
# method = "tags"
|
||||||
|
# keys = ["foo"]
|
||||||
|
# separator = "__"
|
||||||
|
|
||||||
|
## Suffix equals to "_" + measurement's "foo" and "bar"
|
||||||
|
## tag values, separated by "_". If there is no such tags,
|
||||||
|
## their values treated as empty strings.
|
||||||
|
# [outputs.kafka.topic_suffix]
|
||||||
|
# method = "tags"
|
||||||
|
# keys = ["foo", "bar"]
|
||||||
|
# separator = "_"
|
||||||
|
|
||||||
|
## The routing tag specifies a tagkey on the metric whose value is used as
|
||||||
|
## the message key. The message key is used to determine which partition to
|
||||||
|
## send the message to. This tag is prefered over the routing_key option.
|
||||||
|
routing_tag = "host"
|
||||||
|
|
||||||
|
## The routing key is set as the message key and used to determine which
|
||||||
|
## partition to send the message to. This value is only used when no
|
||||||
|
## routing_tag is set or as a fallback when the tag specified in routing tag
|
||||||
|
## is not found.
|
||||||
|
##
|
||||||
|
## If set to "random", a random value will be generated for each message.
|
||||||
|
##
|
||||||
|
## When unset, no message key is added and each message is routed to a random
|
||||||
|
## partition.
|
||||||
|
##
|
||||||
|
## ex: routing_key = "random"
|
||||||
|
## routing_key = "telegraf"
|
||||||
|
# routing_key = ""
|
||||||
|
|
||||||
|
## Compression codec represents the various compression codecs recognized by
|
||||||
|
## Kafka in messages.
|
||||||
|
## 0 : None
|
||||||
|
## 1 : Gzip
|
||||||
|
## 2 : Snappy
|
||||||
|
## 3 : LZ4
|
||||||
|
## 4 : ZSTD
|
||||||
|
# compression_codec = 0
|
||||||
|
|
||||||
|
## Idempotent Writes
|
||||||
|
## If enabled, exactly one copy of each message is written.
|
||||||
|
# idempotent_writes = false
|
||||||
|
|
||||||
|
## RequiredAcks is used in Produce Requests to tell the broker how many
|
||||||
|
## replica acknowledgements it must see before responding
|
||||||
|
## 0 : the producer never waits for an acknowledgement from the broker.
|
||||||
|
## This option provides the lowest latency but the weakest durability
|
||||||
|
## guarantees (some data will be lost when a server fails).
|
||||||
|
## 1 : the producer gets an acknowledgement after the leader replica has
|
||||||
|
## received the data. This option provides better durability as the
|
||||||
|
## client waits until the server acknowledges the request as successful
|
||||||
|
## (only messages that were written to the now-dead leader but not yet
|
||||||
|
## replicated will be lost).
|
||||||
|
## -1: the producer gets an acknowledgement after all in-sync replicas have
|
||||||
|
## received the data. This option provides the best durability, we
|
||||||
|
## guarantee that no messages will be lost as long as at least one in
|
||||||
|
## sync replica remains.
|
||||||
|
# required_acks = -1
|
||||||
|
|
||||||
|
## The maximum number of times to retry sending a metric before failing
|
||||||
|
## until the next flush.
|
||||||
|
# max_retry = 3
|
||||||
|
|
||||||
|
## The maximum permitted size of a message. Should be set equal to or
|
||||||
|
## smaller than the broker's 'message.max.bytes'.
|
||||||
|
# max_message_bytes = 1000000
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Optional SOCKS5 proxy to use when connecting to brokers
|
||||||
|
# socks5_enabled = true
|
||||||
|
# socks5_address = "127.0.0.1:1080"
|
||||||
|
# socks5_username = "alice"
|
||||||
|
# socks5_password = "pass123"
|
||||||
|
|
||||||
|
## Optional SASL Config
|
||||||
|
# sasl_username = "kafka"
|
||||||
|
# sasl_password = "secret"
|
||||||
|
|
||||||
|
## Optional SASL:
|
||||||
|
## one of: OAUTHBEARER, PLAIN, SCRAM-SHA-256, SCRAM-SHA-512, GSSAPI
|
||||||
|
## (defaults to PLAIN)
|
||||||
|
# sasl_mechanism = ""
|
||||||
|
|
||||||
|
## used if sasl_mechanism is GSSAPI (experimental)
|
||||||
|
# sasl_gssapi_service_name = ""
|
||||||
|
# ## One of: KRB5_USER_AUTH and KRB5_KEYTAB_AUTH
|
||||||
|
# sasl_gssapi_auth_type = "KRB5_USER_AUTH"
|
||||||
|
# sasl_gssapi_kerberos_config_path = "/"
|
||||||
|
# sasl_gssapi_realm = "realm"
|
||||||
|
# sasl_gssapi_key_tab_path = ""
|
||||||
|
# sasl_gssapi_disable_pafxfast = false
|
||||||
|
|
||||||
|
## used if sasl_mechanism is OAUTHBEARER (experimental)
|
||||||
|
# sasl_access_token = ""
|
||||||
|
|
||||||
|
## SASL protocol version. When connecting to Azure EventHub set to 0.
|
||||||
|
# sasl_version = 1
|
||||||
|
|
||||||
|
# Disable Kafka metadata full fetch
|
||||||
|
# metadata_full = false
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
@ -0,0 +1,63 @@
|
||||||
|
# Configuration for the AWS Kinesis output.
|
||||||
|
[[outputs.kinesis]]
|
||||||
|
## Amazon REGION of kinesis endpoint.
|
||||||
|
region = "ap-southeast-2"
|
||||||
|
|
||||||
|
## Amazon Credentials
|
||||||
|
## Credentials are loaded in the following order
|
||||||
|
## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||||
|
## 2) Assumed credentials via STS if role_arn is specified
|
||||||
|
## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||||
|
## 4) shared profile from 'profile'
|
||||||
|
## 5) environment variables
|
||||||
|
## 6) shared credentials file
|
||||||
|
## 7) EC2 Instance Profile
|
||||||
|
#access_key = ""
|
||||||
|
#secret_key = ""
|
||||||
|
#token = ""
|
||||||
|
#role_arn = ""
|
||||||
|
#web_identity_token_file = ""
|
||||||
|
#role_session_name = ""
|
||||||
|
#profile = ""
|
||||||
|
#shared_credential_file = ""
|
||||||
|
|
||||||
|
## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
## determined and this option should only be set if you wish to override the
|
||||||
|
## default.
|
||||||
|
## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# endpoint_url = ""
|
||||||
|
|
||||||
|
## Kinesis StreamName must exist prior to starting telegraf.
|
||||||
|
streamname = "StreamName"
|
||||||
|
|
||||||
|
## The partition key can be calculated using one of several methods:
|
||||||
|
##
|
||||||
|
## Use a static value for all writes:
|
||||||
|
# [outputs.kinesis.partition]
|
||||||
|
# method = "static"
|
||||||
|
# key = "howdy"
|
||||||
|
#
|
||||||
|
## Use a random partition key on each write:
|
||||||
|
# [outputs.kinesis.partition]
|
||||||
|
# method = "random"
|
||||||
|
#
|
||||||
|
## Use the measurement name as the partition key:
|
||||||
|
# [outputs.kinesis.partition]
|
||||||
|
# method = "measurement"
|
||||||
|
#
|
||||||
|
## Use the value of a tag for all writes, if the tag is not set the empty
|
||||||
|
## default option will be used. When no default, defaults to "telegraf"
|
||||||
|
# [outputs.kinesis.partition]
|
||||||
|
# method = "tag"
|
||||||
|
# key = "host"
|
||||||
|
# default = "mykey"
|
||||||
|
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
||||||
|
## debug will show upstream aws messages.
|
||||||
|
debug = false
|
||||||
|
|
@ -0,0 +1,16 @@
|
||||||
|
# Configuration for Librato API to send metrics to.
|
||||||
|
[[outputs.librato]]
|
||||||
|
## Librato API Docs
|
||||||
|
## http://dev.librato.com/v1/metrics-authentication
|
||||||
|
## Librato API user
|
||||||
|
api_user = "telegraf@influxdb.com" # required.
|
||||||
|
## Librato API token
|
||||||
|
api_token = "my-secret-token" # required.
|
||||||
|
## Debug
|
||||||
|
# debug = false
|
||||||
|
## Connection timeout.
|
||||||
|
# timeout = "5s"
|
||||||
|
## Output source Template (same as graphite buckets)
|
||||||
|
## see https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md#graphite
|
||||||
|
## This template is used in librato's source (not metric's name)
|
||||||
|
template = "host"
|
||||||
|
|
@ -0,0 +1,23 @@
|
||||||
|
# A plugin that can send metrics over HTTPs to Logz.io
|
||||||
|
[[outputs.logzio]]
|
||||||
|
## Set to true if Logz.io sender checks the disk space before adding metrics to the disk queue.
|
||||||
|
# check_disk_space = true
|
||||||
|
|
||||||
|
## The percent of used file system space at which the sender will stop queueing.
|
||||||
|
## When we will reach that percentage, the file system in which the queue is stored will drop
|
||||||
|
## all new logs until the percentage of used space drops below that threshold.
|
||||||
|
# disk_threshold = 98
|
||||||
|
|
||||||
|
## How often Logz.io sender should drain the queue.
|
||||||
|
## Valid time units are "ns", "us" (or "µs"), "ms", "s", "m", "h".
|
||||||
|
# drain_duration = "3s"
|
||||||
|
|
||||||
|
## Where Logz.io sender should store the queue
|
||||||
|
## queue_dir = Sprintf("%s%s%s%s%d", os.TempDir(), string(os.PathSeparator),
|
||||||
|
## "logzio-buffer", string(os.PathSeparator), time.Now().UnixNano())
|
||||||
|
|
||||||
|
## Logz.io account token
|
||||||
|
token = "your Logz.io token" # required
|
||||||
|
|
||||||
|
## Use your listener URL for your Logz.io account region.
|
||||||
|
# url = "https://listener.logz.io:8071"
|
||||||
|
|
@ -0,0 +1,25 @@
|
||||||
|
# A plugin that can transmit logs to Loki
|
||||||
|
[[outputs.loki]]
|
||||||
|
## The domain of Loki
|
||||||
|
domain = "https://loki.domain.tld"
|
||||||
|
|
||||||
|
## Endpoint to write api
|
||||||
|
# endpoint = "/loki/api/v1/push"
|
||||||
|
|
||||||
|
## Connection timeout, defaults to "5s" if not set.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Basic auth credential
|
||||||
|
# username = "loki"
|
||||||
|
# password = "pass"
|
||||||
|
|
||||||
|
## Additional HTTP headers
|
||||||
|
# http_headers = {"X-Scope-OrgID" = "1"}
|
||||||
|
|
||||||
|
## If the request must be gzip encoded
|
||||||
|
# gzip_request = false
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
|
@ -0,0 +1,34 @@
|
||||||
|
# A plugin that can transmit logs to mongodb
|
||||||
|
[[outputs.mongodb]]
|
||||||
|
# connection string examples for mongodb
|
||||||
|
dsn = "mongodb://localhost:27017"
|
||||||
|
# dsn = "mongodb://mongod1:27017,mongod2:27017,mongod3:27017/admin&replicaSet=myReplSet&w=1"
|
||||||
|
|
||||||
|
# overrides serverSelectionTimeoutMS in dsn if set
|
||||||
|
# timeout = "30s"
|
||||||
|
|
||||||
|
# default authentication, optional
|
||||||
|
# authentication = "NONE"
|
||||||
|
|
||||||
|
# for SCRAM-SHA-256 authentication
|
||||||
|
# authentication = "SCRAM"
|
||||||
|
# username = "root"
|
||||||
|
# password = "***"
|
||||||
|
|
||||||
|
# for x509 certificate authentication
|
||||||
|
# authentication = "X509"
|
||||||
|
# tls_ca = "ca.pem"
|
||||||
|
# tls_key = "client.pem"
|
||||||
|
# # tls_key_pwd = "changeme" # required for encrypted tls_key
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
# database to store measurements and time series collections
|
||||||
|
# database = "telegraf"
|
||||||
|
|
||||||
|
# granularity can be seconds, minutes, or hours.
|
||||||
|
# configuring this value will be based on your input collection frequency.
|
||||||
|
# see https://docs.mongodb.com/manual/core/timeseries-collections/#create-a-time-series-collection
|
||||||
|
# granularity = "seconds"
|
||||||
|
|
||||||
|
# optionally set a TTL to automatically expire documents from the measurement collections.
|
||||||
|
# ttl = "360h"
|
||||||
|
|
@ -0,0 +1,62 @@
|
||||||
|
# Configuration for MQTT server to send metrics to
|
||||||
|
[[outputs.mqtt]]
|
||||||
|
## MQTT Brokers
|
||||||
|
## The list of brokers should only include the hostname or IP address and the
|
||||||
|
## port to the broker. This should follow the format '{host}:{port}'. For
|
||||||
|
## example, "localhost:1883" or "127.0.0.1:8883".
|
||||||
|
servers = ["localhost:1883"]
|
||||||
|
|
||||||
|
## MQTT Topic for Producer Messages
|
||||||
|
## MQTT outputs send metrics to this topic format:
|
||||||
|
## <topic_prefix>/<hostname>/<pluginname>/ (e.g. prefix/web01.example.com/mem)
|
||||||
|
topic_prefix = "telegraf"
|
||||||
|
|
||||||
|
## QoS policy for messages
|
||||||
|
## The mqtt QoS policy for sending messages.
|
||||||
|
## See https://www.ibm.com/support/knowledgecenter/en/SSFKSJ_9.0.0/com.ibm.mq.dev.doc/q029090_.htm
|
||||||
|
## 0 = at most once
|
||||||
|
## 1 = at least once
|
||||||
|
## 2 = exactly once
|
||||||
|
# qos = 2
|
||||||
|
|
||||||
|
## Keep Alive
|
||||||
|
## Defines the maximum length of time that the broker and client may not
|
||||||
|
## communicate. Defaults to 0 which turns the feature off.
|
||||||
|
##
|
||||||
|
## For version v2.0.12 and later mosquitto there is a bug
|
||||||
|
## (see https://github.com/eclipse/mosquitto/issues/2117), which requires
|
||||||
|
## this to be non-zero. As a reference eclipse/paho.mqtt.golang defaults to 30.
|
||||||
|
# keep_alive = 0
|
||||||
|
|
||||||
|
## username and password to connect MQTT server.
|
||||||
|
# username = "telegraf"
|
||||||
|
# password = "metricsmetricsmetricsmetrics"
|
||||||
|
|
||||||
|
## client ID
|
||||||
|
## The unique client id to connect MQTT server. If this parameter is not set
|
||||||
|
## then a random ID is generated.
|
||||||
|
# client_id = ""
|
||||||
|
|
||||||
|
## Timeout for write operations. default: 5s
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## When true, metrics will be sent in one MQTT message per flush. Otherwise,
|
||||||
|
## metrics are written one metric per MQTT message.
|
||||||
|
# batch = false
|
||||||
|
|
||||||
|
## When true, metric will have RETAIN flag set, making broker cache entries until someone
|
||||||
|
## actually reads it
|
||||||
|
# retain = false
|
||||||
|
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
@ -0,0 +1,33 @@
|
||||||
|
# Send telegraf measurements to NATS
|
||||||
|
[[outputs.nats]]
|
||||||
|
## URLs of NATS servers
|
||||||
|
servers = ["nats://localhost:4222"]
|
||||||
|
|
||||||
|
## Optional client name
|
||||||
|
# name = ""
|
||||||
|
|
||||||
|
## Optional credentials
|
||||||
|
# username = ""
|
||||||
|
# password = ""
|
||||||
|
|
||||||
|
## Optional NATS 2.0 and NATS NGS compatible user credentials
|
||||||
|
# credentials = "/etc/telegraf/nats.creds"
|
||||||
|
|
||||||
|
## NATS subject for producer messages
|
||||||
|
subject = "telegraf"
|
||||||
|
|
||||||
|
## Use Transport Layer Security
|
||||||
|
# secure = false
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
@ -0,0 +1,22 @@
|
||||||
|
# Send metrics to New Relic metrics endpoint
|
||||||
|
[[outputs.newrelic]]
|
||||||
|
## The 'insights_key' parameter requires a NR license key.
|
||||||
|
## New Relic recommends you create one
|
||||||
|
## with a convenient name such as TELEGRAF_INSERT_KEY.
|
||||||
|
## reference: https://docs.newrelic.com/docs/apis/intro-apis/new-relic-api-keys/#ingest-license-key
|
||||||
|
# insights_key = "New Relic License Key Here"
|
||||||
|
|
||||||
|
## Prefix to add to add to metric name for easy identification.
|
||||||
|
## This is very useful if your metric names are ambiguous.
|
||||||
|
# metric_prefix = ""
|
||||||
|
|
||||||
|
## Timeout for writes to the New Relic API.
|
||||||
|
# timeout = "15s"
|
||||||
|
|
||||||
|
## HTTP Proxy override. If unset use values from the standard
|
||||||
|
## proxy environment variables to determine proxy, if any.
|
||||||
|
# http_proxy = "http://corporate.proxy:3128"
|
||||||
|
|
||||||
|
## Metric URL override to enable geographic location endpoints.
|
||||||
|
# If not set use values from the standard
|
||||||
|
# metric_url = "https://metric-api.newrelic.com/metric/v1"
|
||||||
|
|
@ -0,0 +1,12 @@
|
||||||
|
# Send telegraf measurements to NSQD
|
||||||
|
[[outputs.nsq]]
|
||||||
|
## Location of nsqd instance listening on TCP
|
||||||
|
server = "localhost:4150"
|
||||||
|
## NSQ topic for producer messages
|
||||||
|
topic = "telegraf"
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
data_format = "influx"
|
||||||
|
|
@ -0,0 +1,33 @@
|
||||||
|
# Send OpenTelemetry metrics over gRPC
|
||||||
|
[[outputs.opentelemetry]]
|
||||||
|
## Override the default (localhost:4317) OpenTelemetry gRPC service
|
||||||
|
## address:port
|
||||||
|
# service_address = "localhost:4317"
|
||||||
|
|
||||||
|
## Override the default (5s) request timeout
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Optional TLS Config.
|
||||||
|
##
|
||||||
|
## Root certificates for verifying server certificates encoded in PEM format.
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
## The public and private keypairs for the client encoded in PEM format.
|
||||||
|
## May contain intermediate certificates.
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS, but skip TLS chain and host verification.
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
## Send the specified TLS server name via SNI.
|
||||||
|
# tls_server_name = "foo.example.com"
|
||||||
|
|
||||||
|
## Override the default (gzip) compression used to send data.
|
||||||
|
## Supports: "gzip", "none"
|
||||||
|
# compression = "gzip"
|
||||||
|
|
||||||
|
## Additional OpenTelemetry resource attributes
|
||||||
|
# [outputs.opentelemetry.attributes]
|
||||||
|
# "service.name" = "demo"
|
||||||
|
|
||||||
|
## Additional gRPC request metadata
|
||||||
|
# [outputs.opentelemetry.headers]
|
||||||
|
# key1 = "value1"
|
||||||
|
|
@ -0,0 +1,26 @@
|
||||||
|
# Configuration for OpenTSDB server to send metrics to
|
||||||
|
[[outputs.opentsdb]]
|
||||||
|
## prefix for metrics keys
|
||||||
|
prefix = "my.specific.prefix."
|
||||||
|
|
||||||
|
## DNS name of the OpenTSDB server
|
||||||
|
## Using "opentsdb.example.com" or "tcp://opentsdb.example.com" will use the
|
||||||
|
## telnet API. "http://opentsdb.example.com" will use the Http API.
|
||||||
|
host = "opentsdb.example.com"
|
||||||
|
|
||||||
|
## Port of the OpenTSDB server
|
||||||
|
port = 4242
|
||||||
|
|
||||||
|
## Number of data points to send to OpenTSDB in Http requests.
|
||||||
|
## Not used with telnet API.
|
||||||
|
http_batch_size = 50
|
||||||
|
|
||||||
|
## URI Path for Http requests to OpenTSDB.
|
||||||
|
## Used in cases where OpenTSDB is located behind a reverse proxy.
|
||||||
|
http_path = "/api/put"
|
||||||
|
|
||||||
|
## Debug true - Prints OpenTSDB communication
|
||||||
|
debug = false
|
||||||
|
|
||||||
|
## Separator separates measurement name from field
|
||||||
|
separator = "_"
|
||||||
|
|
@ -0,0 +1,45 @@
|
||||||
|
# Configuration for the Prometheus client to spawn
|
||||||
|
[[outputs.prometheus_client]]
|
||||||
|
## Address to listen on.
|
||||||
|
listen = ":9273"
|
||||||
|
|
||||||
|
## Metric version controls the mapping from Telegraf metrics into
|
||||||
|
## Prometheus format. When using the prometheus input, use the same value in
|
||||||
|
## both plugins to ensure metrics are round-tripped without modification.
|
||||||
|
##
|
||||||
|
## example: metric_version = 1;
|
||||||
|
## metric_version = 2; recommended version
|
||||||
|
# metric_version = 1
|
||||||
|
|
||||||
|
## Use HTTP Basic Authentication.
|
||||||
|
# basic_username = "Foo"
|
||||||
|
# basic_password = "Bar"
|
||||||
|
|
||||||
|
## If set, the IP Ranges which are allowed to access metrics.
|
||||||
|
## ex: ip_range = ["192.168.0.0/24", "192.168.1.0/30"]
|
||||||
|
# ip_range = []
|
||||||
|
|
||||||
|
## Path to publish the metrics on.
|
||||||
|
# path = "/metrics"
|
||||||
|
|
||||||
|
## Expiration interval for each metric. 0 == no expiration
|
||||||
|
# expiration_interval = "60s"
|
||||||
|
|
||||||
|
## Collectors to enable, valid entries are "gocollector" and "process".
|
||||||
|
## If unset, both are enabled.
|
||||||
|
# collectors_exclude = ["gocollector", "process"]
|
||||||
|
|
||||||
|
## Send string metrics as Prometheus labels.
|
||||||
|
## Unless set to false all string metrics will be sent as labels.
|
||||||
|
# string_as_label = true
|
||||||
|
|
||||||
|
## If set, enable TLS with the given certificate.
|
||||||
|
# tls_cert = "/etc/ssl/telegraf.crt"
|
||||||
|
# tls_key = "/etc/ssl/telegraf.key"
|
||||||
|
|
||||||
|
## Set one or more allowed client CA certificate file names to
|
||||||
|
## enable mutually authenticated TLS connections
|
||||||
|
# tls_allowed_cacerts = ["/etc/telegraf/clientca.pem"]
|
||||||
|
|
||||||
|
## Export metric collection time.
|
||||||
|
# export_timestamp = false
|
||||||
|
|
@ -0,0 +1,32 @@
|
||||||
|
# Configuration for Riemann to send metrics to
|
||||||
|
[[outputs.riemann]]
|
||||||
|
## The full TCP or UDP URL of the Riemann server
|
||||||
|
url = "tcp://localhost:5555"
|
||||||
|
|
||||||
|
## Riemann event TTL, floating-point time in seconds.
|
||||||
|
## Defines how long that an event is considered valid for in Riemann
|
||||||
|
# ttl = 30.0
|
||||||
|
|
||||||
|
## Separator to use between measurement and field name in Riemann service name
|
||||||
|
## This does not have any effect if 'measurement_as_attribute' is set to 'true'
|
||||||
|
separator = "/"
|
||||||
|
|
||||||
|
## Set measurement name as Riemann attribute 'measurement', instead of prepending it to the Riemann service name
|
||||||
|
# measurement_as_attribute = false
|
||||||
|
|
||||||
|
## Send string metrics as Riemann event states.
|
||||||
|
## Unless enabled all string metrics will be ignored
|
||||||
|
# string_as_state = false
|
||||||
|
|
||||||
|
## A list of tag keys whose values get sent as Riemann tags.
|
||||||
|
## If empty, all Telegraf tag values will be sent as tags
|
||||||
|
# tag_keys = ["telegraf","custom_tag"]
|
||||||
|
|
||||||
|
## Additional Riemann tags to send.
|
||||||
|
# tags = ["telegraf-output"]
|
||||||
|
|
||||||
|
## Description for Riemann event
|
||||||
|
# description_text = "metrics collected from telegraf"
|
||||||
|
|
||||||
|
## Riemann client write timeout, defaults to "5s" if not set.
|
||||||
|
# timeout = "5s"
|
||||||
|
|
@ -0,0 +1,8 @@
|
||||||
|
# Configuration for the Riemann server to send metrics to
|
||||||
|
[[outputs.riemann_legacy]]
|
||||||
|
## URL of server
|
||||||
|
url = "localhost:5555"
|
||||||
|
## transport protocol to use either tcp or udp
|
||||||
|
transport = "tcp"
|
||||||
|
## separator to use between input name and field name in Riemann service name
|
||||||
|
separator = " "
|
||||||
|
|
@ -0,0 +1,89 @@
|
||||||
|
# Send aggregate metrics to Sensu Monitor
|
||||||
|
[[outputs.sensu]]
|
||||||
|
## BACKEND API URL is the Sensu Backend API root URL to send metrics to
|
||||||
|
## (protocol, host, and port only). The output plugin will automatically
|
||||||
|
## append the corresponding backend API path
|
||||||
|
## /api/core/v2/namespaces/:entity_namespace/events/:entity_name/:check_name).
|
||||||
|
##
|
||||||
|
## Backend Events API reference:
|
||||||
|
## https://docs.sensu.io/sensu-go/latest/api/events/
|
||||||
|
##
|
||||||
|
## AGENT API URL is the Sensu Agent API root URL to send metrics to
|
||||||
|
## (protocol, host, and port only). The output plugin will automatically
|
||||||
|
## append the correspeonding agent API path (/events).
|
||||||
|
##
|
||||||
|
## Agent API Events API reference:
|
||||||
|
## https://docs.sensu.io/sensu-go/latest/api/events/
|
||||||
|
##
|
||||||
|
## NOTE: if backend_api_url and agent_api_url and api_key are set, the output
|
||||||
|
## plugin will use backend_api_url. If backend_api_url and agent_api_url are
|
||||||
|
## not provided, the output plugin will default to use an agent_api_url of
|
||||||
|
## http://127.0.0.1:3031
|
||||||
|
##
|
||||||
|
# backend_api_url = "http://127.0.0.1:8080"
|
||||||
|
# agent_api_url = "http://127.0.0.1:3031"
|
||||||
|
|
||||||
|
## API KEY is the Sensu Backend API token
|
||||||
|
## Generate a new API token via:
|
||||||
|
##
|
||||||
|
## $ sensuctl cluster-role create telegraf --verb create --resource events,entities
|
||||||
|
## $ sensuctl cluster-role-binding create telegraf --cluster-role telegraf --group telegraf
|
||||||
|
## $ sensuctl user create telegraf --group telegraf --password REDACTED
|
||||||
|
## $ sensuctl api-key grant telegraf
|
||||||
|
##
|
||||||
|
## For more information on Sensu RBAC profiles & API tokens, please visit:
|
||||||
|
## - https://docs.sensu.io/sensu-go/latest/reference/rbac/
|
||||||
|
## - https://docs.sensu.io/sensu-go/latest/reference/apikeys/
|
||||||
|
##
|
||||||
|
# api_key = "${SENSU_API_KEY}"
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Timeout for HTTP message
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## HTTP Content-Encoding for write request body, can be set to "gzip" to
|
||||||
|
## compress body or "identity" to apply no encoding.
|
||||||
|
# content_encoding = "identity"
|
||||||
|
|
||||||
|
## Sensu Event details
|
||||||
|
##
|
||||||
|
## Below are the event details to be sent to Sensu. The main portions of the
|
||||||
|
## event are the check, entity, and metrics specifications. For more information
|
||||||
|
## on Sensu events and its components, please visit:
|
||||||
|
## - Events - https://docs.sensu.io/sensu-go/latest/reference/events
|
||||||
|
## - Checks - https://docs.sensu.io/sensu-go/latest/reference/checks
|
||||||
|
## - Entities - https://docs.sensu.io/sensu-go/latest/reference/entities
|
||||||
|
## - Metrics - https://docs.sensu.io/sensu-go/latest/reference/events#metrics
|
||||||
|
##
|
||||||
|
## Check specification
|
||||||
|
## The check name is the name to give the Sensu check associated with the event
|
||||||
|
## created. This maps to check.metatadata.name in the event.
|
||||||
|
[outputs.sensu.check]
|
||||||
|
name = "telegraf"
|
||||||
|
|
||||||
|
## Entity specification
|
||||||
|
## Configure the entity name and namespace, if necessary. This will be part of
|
||||||
|
## the entity.metadata in the event.
|
||||||
|
##
|
||||||
|
## NOTE: if the output plugin is configured to send events to a
|
||||||
|
## backend_api_url and entity_name is not set, the value returned by
|
||||||
|
## os.Hostname() will be used; if the output plugin is configured to send
|
||||||
|
## events to an agent_api_url, entity_name and entity_namespace are not used.
|
||||||
|
# [outputs.sensu.entity]
|
||||||
|
# name = "server-01"
|
||||||
|
# namespace = "default"
|
||||||
|
|
||||||
|
## Metrics specification
|
||||||
|
## Configure the tags for the metrics that are sent as part of the Sensu event
|
||||||
|
# [outputs.sensu.tags]
|
||||||
|
# source = "telegraf"
|
||||||
|
|
||||||
|
## Configure the handler(s) for processing the provided metrics
|
||||||
|
# [outputs.sensu.metrics]
|
||||||
|
# handlers = ["influxdb","elasticsearch"]
|
||||||
|
|
@ -0,0 +1,17 @@
|
||||||
|
# Send metrics and events to SignalFx
|
||||||
|
[[outputs.signalfx]]
|
||||||
|
## SignalFx Org Access Token
|
||||||
|
access_token = "my-secret-token"
|
||||||
|
|
||||||
|
## The SignalFx realm that your organization resides in
|
||||||
|
signalfx_realm = "us9" # Required if ingest_url is not set
|
||||||
|
|
||||||
|
## You can optionally provide a custom ingest url instead of the
|
||||||
|
## signalfx_realm option above if you are using a gateway or proxy
|
||||||
|
## instance. This option takes precident over signalfx_realm.
|
||||||
|
ingest_url = "https://my-custom-ingest/"
|
||||||
|
|
||||||
|
## Event typed metrics are omitted by default,
|
||||||
|
## If you require an event typed metric you must specify the
|
||||||
|
## metric name in the following list.
|
||||||
|
included_event_names = ["plugin.metric_name"]
|
||||||
|
|
@ -0,0 +1,37 @@
|
||||||
|
# Generic socket writer capable of handling multiple socket types.
|
||||||
|
[[outputs.socket_writer]]
|
||||||
|
## URL to connect to
|
||||||
|
# address = "tcp://127.0.0.1:8094"
|
||||||
|
# address = "tcp://example.com:http"
|
||||||
|
# address = "tcp4://127.0.0.1:8094"
|
||||||
|
# address = "tcp6://127.0.0.1:8094"
|
||||||
|
# address = "tcp6://[2001:db8::1]:8094"
|
||||||
|
# address = "udp://127.0.0.1:8094"
|
||||||
|
# address = "udp4://127.0.0.1:8094"
|
||||||
|
# address = "udp6://127.0.0.1:8094"
|
||||||
|
# address = "unix:///tmp/telegraf.sock"
|
||||||
|
# address = "unixgram:///tmp/telegraf.sock"
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Period between keep alive probes.
|
||||||
|
## Only applies to TCP sockets.
|
||||||
|
## 0 disables keep alive probes.
|
||||||
|
## Defaults to the OS configuration.
|
||||||
|
# keep_alive_period = "5m"
|
||||||
|
|
||||||
|
## Content encoding for message payloads, can be set to "gzip" or to
|
||||||
|
## "identity" to apply no encoding.
|
||||||
|
##
|
||||||
|
# content_encoding = "identity"
|
||||||
|
|
||||||
|
## Data format to generate.
|
||||||
|
## Each data format has its own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
@ -0,0 +1,53 @@
|
||||||
|
# Save metrics to an SQL Database
|
||||||
|
[[outputs.sql]]
|
||||||
|
## Database driver
|
||||||
|
## Valid options: mssql (Microsoft SQL Server), mysql (MySQL), pgx (Postgres),
|
||||||
|
## sqlite (SQLite3), snowflake (snowflake.com) clickhouse (ClickHouse)
|
||||||
|
# driver = ""
|
||||||
|
|
||||||
|
## Data source name
|
||||||
|
## The format of the data source name is different for each database driver.
|
||||||
|
## See the plugin readme for details.
|
||||||
|
# data_source_name = ""
|
||||||
|
|
||||||
|
## Timestamp column name
|
||||||
|
# timestamp_column = "timestamp"
|
||||||
|
|
||||||
|
## Table creation template
|
||||||
|
## Available template variables:
|
||||||
|
## {TABLE} - table name as a quoted identifier
|
||||||
|
## {TABLELITERAL} - table name as a quoted string literal
|
||||||
|
## {COLUMNS} - column definitions (list of quoted identifiers and types)
|
||||||
|
# table_template = "CREATE TABLE {TABLE}({COLUMNS})"
|
||||||
|
|
||||||
|
## Table existence check template
|
||||||
|
## Available template variables:
|
||||||
|
## {TABLE} - tablename as a quoted identifier
|
||||||
|
# table_exists_template = "SELECT 1 FROM {TABLE} LIMIT 1"
|
||||||
|
|
||||||
|
## Initialization SQL
|
||||||
|
# init_sql = ""
|
||||||
|
|
||||||
|
## Metric type to SQL type conversion
|
||||||
|
## The values on the left are the data types Telegraf has and the values on
|
||||||
|
## the right are the data types Telegraf will use when sending to a database.
|
||||||
|
##
|
||||||
|
## The database values used must be data types the destination database
|
||||||
|
## understands. It is up to the user to ensure that the selected data type is
|
||||||
|
## available in the database they are using. Refer to your database
|
||||||
|
## documentation for what data types are available and supported.
|
||||||
|
#[outputs.sql.convert]
|
||||||
|
# integer = "INT"
|
||||||
|
# real = "DOUBLE"
|
||||||
|
# text = "TEXT"
|
||||||
|
# timestamp = "TIMESTAMP"
|
||||||
|
# defaultvalue = "TEXT"
|
||||||
|
# unsigned = "UNSIGNED"
|
||||||
|
# bool = "BOOL"
|
||||||
|
|
||||||
|
## This setting controls the behavior of the unsigned value. By default the
|
||||||
|
## setting will take the integer value and append the unsigned value to it. The other
|
||||||
|
## option is "literal", which will use the actual value the user provides to
|
||||||
|
## the unsigned option. This is useful for a database like ClickHouse where
|
||||||
|
## the unsigned value should use a value like "uint64".
|
||||||
|
# conversion_style = "unsigned_suffix"
|
||||||
|
|
@ -0,0 +1,16 @@
|
||||||
|
# Configuration for Google Cloud Stackdriver to send metrics to
|
||||||
|
[[outputs.stackdriver]]
|
||||||
|
## GCP Project
|
||||||
|
project = "erudite-bloom-151019"
|
||||||
|
|
||||||
|
## The namespace for the metric descriptor
|
||||||
|
namespace = "telegraf"
|
||||||
|
|
||||||
|
## Custom resource type
|
||||||
|
# resource_type = "generic_node"
|
||||||
|
|
||||||
|
## Additional resource labels
|
||||||
|
# [outputs.stackdriver.resource_labels]
|
||||||
|
# node_id = "$HOSTNAME"
|
||||||
|
# namespace = "myapp"
|
||||||
|
# location = "eu-north0"
|
||||||
|
|
@ -0,0 +1,51 @@
|
||||||
|
# A plugin that can send metrics to Sumo Logic HTTP metric collector.
|
||||||
|
[[outputs.sumologic]]
|
||||||
|
## Unique URL generated for your HTTP Metrics Source.
|
||||||
|
## This is the address to send metrics to.
|
||||||
|
# url = "https://events.sumologic.net/receiver/v1/http/<UniqueHTTPCollectorCode>"
|
||||||
|
|
||||||
|
## Data format to be used for sending metrics.
|
||||||
|
## This will set the "Content-Type" header accordingly.
|
||||||
|
## Currently supported formats:
|
||||||
|
## * graphite - for Content-Type of application/vnd.sumologic.graphite
|
||||||
|
## * carbon2 - for Content-Type of application/vnd.sumologic.carbon2
|
||||||
|
## * prometheus - for Content-Type of application/vnd.sumologic.prometheus
|
||||||
|
##
|
||||||
|
## More information can be found at:
|
||||||
|
## https://help.sumologic.com/03Send-Data/Sources/02Sources-for-Hosted-Collectors/HTTP-Source/Upload-Metrics-to-an-HTTP-Source#content-type-headers-for-metrics
|
||||||
|
##
|
||||||
|
## NOTE:
|
||||||
|
## When unset, telegraf will by default use the influx serializer which is currently unsupported
|
||||||
|
## in HTTP Source.
|
||||||
|
data_format = "carbon2"
|
||||||
|
|
||||||
|
## Timeout used for HTTP request
|
||||||
|
# timeout = "5s"
|
||||||
|
|
||||||
|
## Max HTTP request body size in bytes before compression (if applied).
|
||||||
|
## By default 1MB is recommended.
|
||||||
|
## NOTE:
|
||||||
|
## Bear in mind that in some serializer a metric even though serialized to multiple
|
||||||
|
## lines cannot be split any further so setting this very low might not work
|
||||||
|
## as expected.
|
||||||
|
# max_request_body_size = 1000000
|
||||||
|
|
||||||
|
## Additional, Sumo specific options.
|
||||||
|
## Full list can be found here:
|
||||||
|
## https://help.sumologic.com/03Send-Data/Sources/02Sources-for-Hosted-Collectors/HTTP-Source/Upload-Metrics-to-an-HTTP-Source#supported-http-headers
|
||||||
|
|
||||||
|
## Desired source name.
|
||||||
|
## Useful if you want to override the source name configured for the source.
|
||||||
|
# source_name = ""
|
||||||
|
|
||||||
|
## Desired host name.
|
||||||
|
## Useful if you want to override the source host configured for the source.
|
||||||
|
# source_host = ""
|
||||||
|
|
||||||
|
## Desired source category.
|
||||||
|
## Useful if you want to override the source category configured for the source.
|
||||||
|
# source_category = ""
|
||||||
|
|
||||||
|
## Comma-separated key=value list of dimensions to apply to every metric.
|
||||||
|
## Custom dimensions will allow you to query your metrics at a more granular level.
|
||||||
|
# dimensions = ""
|
||||||
|
|
@ -0,0 +1,76 @@
|
||||||
|
# Configuration for Syslog server to send metrics to
|
||||||
|
[[outputs.syslog]]
|
||||||
|
## URL to connect to
|
||||||
|
## ex: address = "tcp://127.0.0.1:8094"
|
||||||
|
## ex: address = "tcp4://127.0.0.1:8094"
|
||||||
|
## ex: address = "tcp6://127.0.0.1:8094"
|
||||||
|
## ex: address = "tcp6://[2001:db8::1]:8094"
|
||||||
|
## ex: address = "udp://127.0.0.1:8094"
|
||||||
|
## ex: address = "udp4://127.0.0.1:8094"
|
||||||
|
## ex: address = "udp6://127.0.0.1:8094"
|
||||||
|
address = "tcp://127.0.0.1:8094"
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Period between keep alive probes.
|
||||||
|
## Only applies to TCP sockets.
|
||||||
|
## 0 disables keep alive probes.
|
||||||
|
## Defaults to the OS configuration.
|
||||||
|
# keep_alive_period = "5m"
|
||||||
|
|
||||||
|
## The framing technique with which it is expected that messages are
|
||||||
|
## transported (default = "octet-counting"). Whether the messages come
|
||||||
|
## using the octect-counting (RFC5425#section-4.3.1, RFC6587#section-3.4.1),
|
||||||
|
## or the non-transparent framing technique (RFC6587#section-3.4.2). Must
|
||||||
|
## be one of "octet-counting", "non-transparent".
|
||||||
|
# framing = "octet-counting"
|
||||||
|
|
||||||
|
## The trailer to be expected in case of non-transparent framing (default = "LF").
|
||||||
|
## Must be one of "LF", or "NUL".
|
||||||
|
# trailer = "LF"
|
||||||
|
|
||||||
|
## SD-PARAMs settings
|
||||||
|
## Syslog messages can contain key/value pairs within zero or more
|
||||||
|
## structured data sections. For each unrecognized metric tag/field a
|
||||||
|
## SD-PARAMS is created.
|
||||||
|
##
|
||||||
|
## Example:
|
||||||
|
## [[outputs.syslog]]
|
||||||
|
## sdparam_separator = "_"
|
||||||
|
## default_sdid = "default@32473"
|
||||||
|
## sdids = ["foo@123", "bar@456"]
|
||||||
|
##
|
||||||
|
## input => xyzzy,x=y foo@123_value=42,bar@456_value2=84,something_else=1
|
||||||
|
## output (structured data only) => [foo@123 value=42][bar@456 value2=84][default@32473 something_else=1 x=y]
|
||||||
|
|
||||||
|
## SD-PARAMs separator between the sdid and tag/field key (default = "_")
|
||||||
|
# sdparam_separator = "_"
|
||||||
|
|
||||||
|
## Default sdid used for tags/fields that don't contain a prefix defined in
|
||||||
|
## the explicit sdids setting below If no default is specified, no SD-PARAMs
|
||||||
|
## will be used for unrecognized field.
|
||||||
|
# default_sdid = "default@32473"
|
||||||
|
|
||||||
|
## List of explicit prefixes to extract from tag/field keys and use as the
|
||||||
|
## SDID, if they match (see above example for more details):
|
||||||
|
# sdids = ["foo@123", "bar@456"]
|
||||||
|
|
||||||
|
## Default severity value. Severity and Facility are used to calculate the
|
||||||
|
## message PRI value (RFC5424#section-6.2.1). Used when no metric field
|
||||||
|
## with key "severity_code" is defined. If unset, 5 (notice) is the default
|
||||||
|
# default_severity_code = 5
|
||||||
|
|
||||||
|
## Default facility value. Facility and Severity are used to calculate the
|
||||||
|
## message PRI value (RFC5424#section-6.2.1). Used when no metric field with
|
||||||
|
## key "facility_code" is defined. If unset, 1 (user-level) is the default
|
||||||
|
# default_facility_code = 1
|
||||||
|
|
||||||
|
## Default APP-NAME value (RFC5424#section-6.2.5)
|
||||||
|
## Used when no metric tag with key "appname" is defined.
|
||||||
|
## If unset, "Telegraf" is the default
|
||||||
|
# default_appname = "Telegraf"
|
||||||
|
|
@ -0,0 +1,117 @@
|
||||||
|
# Configuration for sending metrics to Amazon Timestream.
|
||||||
|
[[outputs.timestream]]
|
||||||
|
## Amazon Region
|
||||||
|
region = "us-east-1"
|
||||||
|
|
||||||
|
## Amazon Credentials
|
||||||
|
## Credentials are loaded in the following order
|
||||||
|
## 1) Web identity provider credentials via STS if role_arn and web_identity_token_file are specified
|
||||||
|
## 2) Assumed credentials via STS if role_arn is specified
|
||||||
|
## 3) explicit credentials from 'access_key' and 'secret_key'
|
||||||
|
## 4) shared profile from 'profile'
|
||||||
|
## 5) environment variables
|
||||||
|
## 6) shared credentials file
|
||||||
|
## 7) EC2 Instance Profile
|
||||||
|
#access_key = ""
|
||||||
|
#secret_key = ""
|
||||||
|
#token = ""
|
||||||
|
#role_arn = ""
|
||||||
|
#web_identity_token_file = ""
|
||||||
|
#role_session_name = ""
|
||||||
|
#profile = ""
|
||||||
|
#shared_credential_file = ""
|
||||||
|
|
||||||
|
## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
## determined and this option should only be set if you wish to override the
|
||||||
|
## default.
|
||||||
|
## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# endpoint_url = ""
|
||||||
|
|
||||||
|
## Timestream database where the metrics will be inserted.
|
||||||
|
## The database must exist prior to starting Telegraf.
|
||||||
|
database_name = "yourDatabaseNameHere"
|
||||||
|
|
||||||
|
## Specifies if the plugin should describe the Timestream database upon starting
|
||||||
|
## to validate if it has access necessary permissions, connection, etc., as a safety check.
|
||||||
|
## If the describe operation fails, the plugin will not start
|
||||||
|
## and therefore the Telegraf agent will not start.
|
||||||
|
describe_database_on_start = false
|
||||||
|
|
||||||
|
## The mapping mode specifies how Telegraf records are represented in Timestream.
|
||||||
|
## Valid values are: single-table, multi-table.
|
||||||
|
## For example, consider the following data in line protocol format:
|
||||||
|
## weather,location=us-midwest,season=summer temperature=82,humidity=71 1465839830100400200
|
||||||
|
## airquality,location=us-west no2=5,pm25=16 1465839830100400200
|
||||||
|
## where weather and airquality are the measurement names, location and season are tags,
|
||||||
|
## and temperature, humidity, no2, pm25 are fields.
|
||||||
|
## In multi-table mode:
|
||||||
|
## - first line will be ingested to table named weather
|
||||||
|
## - second line will be ingested to table named airquality
|
||||||
|
## - the tags will be represented as dimensions
|
||||||
|
## - first table (weather) will have two records:
|
||||||
|
## one with measurement name equals to temperature,
|
||||||
|
## another with measurement name equals to humidity
|
||||||
|
## - second table (airquality) will have two records:
|
||||||
|
## one with measurement name equals to no2,
|
||||||
|
## another with measurement name equals to pm25
|
||||||
|
## - the Timestream tables from the example will look like this:
|
||||||
|
## TABLE "weather":
|
||||||
|
## time | location | season | measure_name | measure_value::bigint
|
||||||
|
## 2016-06-13 17:43:50 | us-midwest | summer | temperature | 82
|
||||||
|
## 2016-06-13 17:43:50 | us-midwest | summer | humidity | 71
|
||||||
|
## TABLE "airquality":
|
||||||
|
## time | location | measure_name | measure_value::bigint
|
||||||
|
## 2016-06-13 17:43:50 | us-west | no2 | 5
|
||||||
|
## 2016-06-13 17:43:50 | us-west | pm25 | 16
|
||||||
|
## In single-table mode:
|
||||||
|
## - the data will be ingested to a single table, which name will be valueOf(single_table_name)
|
||||||
|
## - measurement name will stored in dimension named valueOf(single_table_dimension_name_for_telegraf_measurement_name)
|
||||||
|
## - location and season will be represented as dimensions
|
||||||
|
## - temperature, humidity, no2, pm25 will be represented as measurement name
|
||||||
|
## - the Timestream table from the example will look like this:
|
||||||
|
## Assuming:
|
||||||
|
## - single_table_name = "my_readings"
|
||||||
|
## - single_table_dimension_name_for_telegraf_measurement_name = "namespace"
|
||||||
|
## TABLE "my_readings":
|
||||||
|
## time | location | season | namespace | measure_name | measure_value::bigint
|
||||||
|
## 2016-06-13 17:43:50 | us-midwest | summer | weather | temperature | 82
|
||||||
|
## 2016-06-13 17:43:50 | us-midwest | summer | weather | humidity | 71
|
||||||
|
## 2016-06-13 17:43:50 | us-west | NULL | airquality | no2 | 5
|
||||||
|
## 2016-06-13 17:43:50 | us-west | NULL | airquality | pm25 | 16
|
||||||
|
## In most cases, using multi-table mapping mode is recommended.
|
||||||
|
## However, you can consider using single-table in situations when you have thousands of measurement names.
|
||||||
|
mapping_mode = "multi-table"
|
||||||
|
|
||||||
|
## Only valid and required for mapping_mode = "single-table"
|
||||||
|
## Specifies the Timestream table where the metrics will be uploaded.
|
||||||
|
# single_table_name = "yourTableNameHere"
|
||||||
|
|
||||||
|
## Only valid and required for mapping_mode = "single-table"
|
||||||
|
## Describes what will be the Timestream dimension name for the Telegraf
|
||||||
|
## measurement name.
|
||||||
|
# single_table_dimension_name_for_telegraf_measurement_name = "namespace"
|
||||||
|
|
||||||
|
## Specifies if the plugin should create the table, if the table do not exist.
|
||||||
|
## The plugin writes the data without prior checking if the table exists.
|
||||||
|
## When the table does not exist, the error returned from Timestream will cause
|
||||||
|
## the plugin to create the table, if this parameter is set to true.
|
||||||
|
create_table_if_not_exists = true
|
||||||
|
|
||||||
|
## Only valid and required if create_table_if_not_exists = true
|
||||||
|
## Specifies the Timestream table magnetic store retention period in days.
|
||||||
|
## Check Timestream documentation for more details.
|
||||||
|
create_table_magnetic_store_retention_period_in_days = 365
|
||||||
|
|
||||||
|
## Only valid and required if create_table_if_not_exists = true
|
||||||
|
## Specifies the Timestream table memory store retention period in hours.
|
||||||
|
## Check Timestream documentation for more details.
|
||||||
|
create_table_memory_store_retention_period_in_hours = 24
|
||||||
|
|
||||||
|
## Only valid and optional if create_table_if_not_exists = true
|
||||||
|
## Specifies the Timestream table tags.
|
||||||
|
## Check Timestream documentation for more details
|
||||||
|
# create_table_tags = { "foo" = "bar", "environment" = "dev"}
|
||||||
|
|
||||||
|
## Specify the maximum number of parallel go routines to ingest/write data
|
||||||
|
## If not specified, defaulted to 1 go routines
|
||||||
|
max_write_go_routines = 25
|
||||||
|
|
@ -0,0 +1,26 @@
|
||||||
|
# Write metrics to Warp 10
|
||||||
|
[[outputs.warp10]]
|
||||||
|
# Prefix to add to the measurement.
|
||||||
|
prefix = "telegraf."
|
||||||
|
|
||||||
|
# URL of the Warp 10 server
|
||||||
|
warp_url = "http://localhost:8080"
|
||||||
|
|
||||||
|
# Write token to access your app on warp 10
|
||||||
|
token = "Token"
|
||||||
|
|
||||||
|
# Warp 10 query timeout
|
||||||
|
# timeout = "15s"
|
||||||
|
|
||||||
|
## Print Warp 10 error body
|
||||||
|
# print_error_body = false
|
||||||
|
|
||||||
|
## Max string error size
|
||||||
|
# max_string_error_size = 511
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
@ -0,0 +1,51 @@
|
||||||
|
# Configuration for Wavefront server to send metrics to
|
||||||
|
[[outputs.wavefront]]
|
||||||
|
## Url for Wavefront Direct Ingestion. For Wavefront Proxy Ingestion, see
|
||||||
|
## the 'host' and 'port' options below.
|
||||||
|
url = "https://metrics.wavefront.com"
|
||||||
|
|
||||||
|
## Authentication Token for Wavefront. Only required if using Direct Ingestion
|
||||||
|
#token = "DUMMY_TOKEN"
|
||||||
|
|
||||||
|
## DNS name of the wavefront proxy server. Do not use if url is specified
|
||||||
|
#host = "wavefront.example.com"
|
||||||
|
|
||||||
|
## Port that the Wavefront proxy server listens on. Do not use if url is specified
|
||||||
|
#port = 2878
|
||||||
|
|
||||||
|
## prefix for metrics keys
|
||||||
|
#prefix = "my.specific.prefix."
|
||||||
|
|
||||||
|
## whether to use "value" for name of simple fields. default is false
|
||||||
|
#simple_fields = false
|
||||||
|
|
||||||
|
## character to use between metric and field name. default is . (dot)
|
||||||
|
#metric_separator = "."
|
||||||
|
|
||||||
|
## Convert metric name paths to use metricSeparator character
|
||||||
|
## When true will convert all _ (underscore) characters in final metric name. default is true
|
||||||
|
#convert_paths = true
|
||||||
|
|
||||||
|
## Use Strict rules to sanitize metric and tag names from invalid characters
|
||||||
|
## When enabled forward slash (/) and comma (,) will be accepted
|
||||||
|
#use_strict = false
|
||||||
|
|
||||||
|
## Use Regex to sanitize metric and tag names from invalid characters
|
||||||
|
## Regex is more thorough, but significantly slower. default is false
|
||||||
|
#use_regex = false
|
||||||
|
|
||||||
|
## point tags to use as the source name for Wavefront (if none found, host will be used)
|
||||||
|
#source_override = ["hostname", "address", "agent_host", "node_host"]
|
||||||
|
|
||||||
|
## whether to convert boolean values to numeric values, with false -> 0.0 and true -> 1.0. default is true
|
||||||
|
#convert_bool = true
|
||||||
|
|
||||||
|
## Truncate metric tags to a total of 254 characters for the tag name value. Wavefront will reject any
|
||||||
|
## data point exceeding this limit if not truncated. Defaults to 'false' to provide backwards compatibility.
|
||||||
|
#truncate_tags = false
|
||||||
|
|
||||||
|
## Flush the internal buffers after each batch. This effectively bypasses the background sending of metrics
|
||||||
|
## normally done by the Wavefront SDK. This can be used if you are experiencing buffer overruns. The sending
|
||||||
|
## of metrics will block for a longer time, but this will be handled gracefully by the internal buffering in
|
||||||
|
## Telegraf.
|
||||||
|
#immediate_flush = true
|
||||||
|
|
@ -0,0 +1,35 @@
|
||||||
|
# A plugin that can transmit metrics over WebSocket.
|
||||||
|
[[outputs.websocket]]
|
||||||
|
## URL is the address to send metrics to. Make sure ws or wss scheme is used.
|
||||||
|
url = "ws://127.0.0.1:3000/telegraf"
|
||||||
|
|
||||||
|
## Timeouts (make sure read_timeout is larger than server ping interval or set to zero).
|
||||||
|
# connect_timeout = "30s"
|
||||||
|
# write_timeout = "30s"
|
||||||
|
# read_timeout = "30s"
|
||||||
|
|
||||||
|
## Optionally turn on using text data frames (binary by default).
|
||||||
|
# use_text_frames = false
|
||||||
|
|
||||||
|
## Optional TLS Config
|
||||||
|
# tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# tls_key = "/etc/telegraf/key.pem"
|
||||||
|
## Use TLS but skip chain & host verification
|
||||||
|
# insecure_skip_verify = false
|
||||||
|
|
||||||
|
## Optional SOCKS5 proxy to use
|
||||||
|
# socks5_enabled = true
|
||||||
|
# socks5_address = "127.0.0.1:1080"
|
||||||
|
# socks5_username = "alice"
|
||||||
|
# socks5_password = "pass123"
|
||||||
|
|
||||||
|
## Data format to output.
|
||||||
|
## Each data format has it's own unique set of configuration options, read
|
||||||
|
## more about them here:
|
||||||
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
||||||
|
## Additional HTTP Upgrade headers
|
||||||
|
# [outputs.websocket.headers]
|
||||||
|
# Authorization = "Bearer <TOKEN>"
|
||||||
|
|
@ -0,0 +1,10 @@
|
||||||
|
# Send aggregated metrics to Yandex.Cloud Monitoring
|
||||||
|
[[outputs.yandex_cloud_monitoring]]
|
||||||
|
## Timeout for HTTP writes.
|
||||||
|
# timeout = "20s"
|
||||||
|
|
||||||
|
## Yandex.Cloud monitoring API endpoint. Normally should not be changed
|
||||||
|
# endpoint_url = "https://monitoring.api.cloud.yandex.net/monitoring/v2/data/write"
|
||||||
|
|
||||||
|
## All user metrics should be sent with "custom" service specified. Normally should not be changed
|
||||||
|
# service = "custom"
|
||||||
Loading…
Reference in New Issue