933 lines
59 KiB
Markdown
933 lines
59 KiB
Markdown
# VMware vSphere Input Plugin
|
|
|
|
The VMware vSphere plugin uses the vSphere API to gather metrics from multiple
|
|
vCenter servers.
|
|
|
|
* Clusters
|
|
* Hosts
|
|
* Resource Pools
|
|
* VMs
|
|
* Datastores
|
|
* vSAN
|
|
|
|
## Supported versions of vSphere
|
|
|
|
This plugin supports vSphere version 6.5, 6.7, 7.0 and 8.0.
|
|
It may work with versions 5.1, 5.5 and 6.0, but neither are
|
|
officially supported.
|
|
|
|
Compatibility information is available from the govmomi project
|
|
[here](https://github.com/vmware/govmomi/tree/v0.26.0#compatibility)
|
|
|
|
## Global configuration options <!-- @/docs/includes/plugin_config.md -->
|
|
|
|
In addition to the plugin-specific configuration settings, plugins support
|
|
additional global and plugin configuration settings. These settings are used to
|
|
modify metrics, tags, and field or create aliases and configure ordering, etc.
|
|
See the [CONFIGURATION.md][CONFIGURATION.md] for more details.
|
|
|
|
[CONFIGURATION.md]: ../../../docs/CONFIGURATION.md#plugins
|
|
|
|
## Secret-store support
|
|
|
|
This plugin supports secrets from secret-stores for the `username` and
|
|
`password` option.
|
|
See the [secret-store documentation][SECRETSTORE] for more details on how
|
|
to use them.
|
|
|
|
[SECRETSTORE]: ../../../docs/CONFIGURATION.md#secret-store-secrets
|
|
|
|
## Configuration
|
|
|
|
```toml @sample.conf
|
|
# Read metrics from one or many vCenters
|
|
[[inputs.vsphere]]
|
|
## List of vCenter URLs to be monitored. These three lines must be uncommented
|
|
## and edited for the plugin to work.
|
|
vcenters = [ "https://vcenter.local/sdk" ]
|
|
username = "user@corp.local"
|
|
password = "secret"
|
|
|
|
## VMs
|
|
## Typical VM metrics (if omitted or empty, all metrics are collected)
|
|
# vm_include = [ "/*/vm/**"] # Inventory path to VMs to collect (by default all are collected)
|
|
# vm_exclude = [] # Inventory paths to exclude
|
|
vm_metric_include = [
|
|
"cpu.demand.average",
|
|
"cpu.idle.summation",
|
|
"cpu.latency.average",
|
|
"cpu.readiness.average",
|
|
"cpu.ready.summation",
|
|
"cpu.run.summation",
|
|
"cpu.usagemhz.average",
|
|
"cpu.used.summation",
|
|
"cpu.wait.summation",
|
|
"mem.active.average",
|
|
"mem.granted.average",
|
|
"mem.latency.average",
|
|
"mem.swapin.average",
|
|
"mem.swapinRate.average",
|
|
"mem.swapout.average",
|
|
"mem.swapoutRate.average",
|
|
"mem.usage.average",
|
|
"mem.vmmemctl.average",
|
|
"net.bytesRx.average",
|
|
"net.bytesTx.average",
|
|
"net.droppedRx.summation",
|
|
"net.droppedTx.summation",
|
|
"net.usage.average",
|
|
"power.power.average",
|
|
"virtualDisk.numberReadAveraged.average",
|
|
"virtualDisk.numberWriteAveraged.average",
|
|
"virtualDisk.read.average",
|
|
"virtualDisk.readOIO.latest",
|
|
"virtualDisk.throughput.usage.average",
|
|
"virtualDisk.totalReadLatency.average",
|
|
"virtualDisk.totalWriteLatency.average",
|
|
"virtualDisk.write.average",
|
|
"virtualDisk.writeOIO.latest",
|
|
"sys.uptime.latest",
|
|
]
|
|
# vm_metric_exclude = [] ## Nothing is excluded by default
|
|
# vm_instances = true ## true by default
|
|
|
|
## Hosts
|
|
## Typical host metrics (if omitted or empty, all metrics are collected)
|
|
# host_include = [ "/*/host/**"] # Inventory path to hosts to collect (by default all are collected)
|
|
# host_exclude [] # Inventory paths to exclude
|
|
host_metric_include = [
|
|
"cpu.coreUtilization.average",
|
|
"cpu.costop.summation",
|
|
"cpu.demand.average",
|
|
"cpu.idle.summation",
|
|
"cpu.latency.average",
|
|
"cpu.readiness.average",
|
|
"cpu.ready.summation",
|
|
"cpu.swapwait.summation",
|
|
"cpu.usage.average",
|
|
"cpu.usagemhz.average",
|
|
"cpu.used.summation",
|
|
"cpu.utilization.average",
|
|
"cpu.wait.summation",
|
|
"disk.deviceReadLatency.average",
|
|
"disk.deviceWriteLatency.average",
|
|
"disk.kernelReadLatency.average",
|
|
"disk.kernelWriteLatency.average",
|
|
"disk.numberReadAveraged.average",
|
|
"disk.numberWriteAveraged.average",
|
|
"disk.read.average",
|
|
"disk.totalReadLatency.average",
|
|
"disk.totalWriteLatency.average",
|
|
"disk.write.average",
|
|
"mem.active.average",
|
|
"mem.latency.average",
|
|
"mem.state.latest",
|
|
"mem.swapin.average",
|
|
"mem.swapinRate.average",
|
|
"mem.swapout.average",
|
|
"mem.swapoutRate.average",
|
|
"mem.totalCapacity.average",
|
|
"mem.usage.average",
|
|
"mem.vmmemctl.average",
|
|
"net.bytesRx.average",
|
|
"net.bytesTx.average",
|
|
"net.droppedRx.summation",
|
|
"net.droppedTx.summation",
|
|
"net.errorsRx.summation",
|
|
"net.errorsTx.summation",
|
|
"net.usage.average",
|
|
"power.power.average",
|
|
"storageAdapter.numberReadAveraged.average",
|
|
"storageAdapter.numberWriteAveraged.average",
|
|
"storageAdapter.read.average",
|
|
"storageAdapter.write.average",
|
|
"sys.uptime.latest",
|
|
]
|
|
## Collect IP addresses? Valid values are "ipv4" and "ipv6"
|
|
# ip_addresses = ["ipv6", "ipv4" ]
|
|
|
|
# host_metric_exclude = [] ## Nothing excluded by default
|
|
# host_instances = true ## true by default
|
|
|
|
|
|
## Clusters
|
|
# cluster_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
|
|
# cluster_exclude = [] # Inventory paths to exclude
|
|
# cluster_metric_include = [] ## if omitted or empty, all metrics are collected
|
|
# cluster_metric_exclude = [] ## Nothing excluded by default
|
|
# cluster_instances = false ## false by default
|
|
|
|
## Resource Pools
|
|
# resource_pool_include = [ "/*/host/**"] # Inventory path to resource pools to collect (by default all are collected)
|
|
# resource_pool_exclude = [] # Inventory paths to exclude
|
|
# resource_pool_metric_include = [] ## if omitted or empty, all metrics are collected
|
|
# resource_pool_metric_exclude = [] ## Nothing excluded by default
|
|
# resource_pool_instances = false ## false by default
|
|
|
|
## Datastores
|
|
# datastore_include = [ "/*/datastore/**"] # Inventory path to datastores to collect (by default all are collected)
|
|
# datastore_exclude = [] # Inventory paths to exclude
|
|
# datastore_metric_include = [] ## if omitted or empty, all metrics are collected
|
|
# datastore_metric_exclude = [] ## Nothing excluded by default
|
|
# datastore_instances = false ## false by default
|
|
|
|
## Datacenters
|
|
# datacenter_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
|
|
# datacenter_exclude = [] # Inventory paths to exclude
|
|
datacenter_metric_include = [] ## if omitted or empty, all metrics are collected
|
|
datacenter_metric_exclude = [ "*" ] ## Datacenters are not collected by default.
|
|
# datacenter_instances = false ## false by default
|
|
|
|
## VSAN
|
|
# vsan_metric_include = [] ## if omitted or empty, all metrics are collected
|
|
# vsan_metric_exclude = [ "*" ] ## vSAN are not collected by default.
|
|
## Whether to skip verifying vSAN metrics against the ones from GetSupportedEntityTypes API.
|
|
# vsan_metric_skip_verify = false ## false by default.
|
|
|
|
## Interval for sampling vSAN performance metrics, can be reduced down to
|
|
## 30 seconds for vSAN 8 U1.
|
|
# vsan_interval = "5m"
|
|
|
|
## Plugin Settings
|
|
## separator character to use for measurement and field names (default: "_")
|
|
# separator = "_"
|
|
|
|
## number of objects to retrieve per query for realtime resources (vms and hosts)
|
|
## set to 64 for vCenter 5.5 and 6.0 (default: 256)
|
|
# max_query_objects = 256
|
|
|
|
## number of metrics to retrieve per query for non-realtime resources (clusters and datastores)
|
|
## set to 64 for vCenter 5.5 and 6.0 (default: 256)
|
|
# max_query_metrics = 256
|
|
|
|
## number of go routines to use for collection and discovery of objects and metrics
|
|
# collect_concurrency = 1
|
|
# discover_concurrency = 1
|
|
|
|
## the interval before (re)discovering objects subject to metrics collection (default: 300s)
|
|
# object_discovery_interval = "300s"
|
|
|
|
## timeout applies to any of the api request made to vcenter
|
|
# timeout = "60s"
|
|
|
|
## When set to true, all samples are sent as integers. This makes the output
|
|
## data types backwards compatible with Telegraf 1.9 or lower. Normally all
|
|
## samples from vCenter, with the exception of percentages, are integer
|
|
## values, but under some conditions, some averaging takes place internally in
|
|
## the plugin. Setting this flag to "false" will send values as floats to
|
|
## preserve the full precision when averaging takes place.
|
|
# use_int_samples = true
|
|
|
|
## Custom attributes from vCenter can be very useful for queries in order to slice the
|
|
## metrics along different dimension and for forming ad-hoc relationships. They are disabled
|
|
## by default, since they can add a considerable amount of tags to the resulting metrics. To
|
|
## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
|
|
## to select the attributes you want to include.
|
|
## By default, since they can add a considerable amount of tags to the resulting metrics. To
|
|
## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
|
|
## to select the attributes you want to include.
|
|
# custom_attribute_include = []
|
|
# custom_attribute_exclude = ["*"]
|
|
|
|
## The number of vSphere 5 minute metric collection cycles to look back for non-realtime metrics. In
|
|
## some versions (6.7, 7.0 and possible more), certain metrics, such as cluster metrics, may be reported
|
|
## with a significant delay (>30min). If this happens, try increasing this number. Please note that increasing
|
|
## it too much may cause performance issues.
|
|
# metric_lookback = 3
|
|
|
|
## Optional SSL Config
|
|
# ssl_ca = "/path/to/cafile"
|
|
# ssl_cert = "/path/to/certfile"
|
|
# ssl_key = "/path/to/keyfile"
|
|
## Use SSL but skip chain & host verification
|
|
# insecure_skip_verify = false
|
|
|
|
## The Historical Interval value must match EXACTLY the interval in the daily
|
|
# "Interval Duration" found on the VCenter server under Configure > General > Statistics > Statistic intervals
|
|
# historical_interval = "5m"
|
|
|
|
## Specifies plugin behavior regarding disconnected servers
|
|
## Available choices :
|
|
## - error: telegraf will return an error on startup if one the servers is unreachable
|
|
## - ignore: telegraf will ignore unreachable servers on both startup and gather
|
|
# disconnected_servers_behavior = "error"
|
|
|
|
## HTTP Proxy support
|
|
# use_system_proxy = true
|
|
# http_proxy_url = ""
|
|
```
|
|
|
|
NOTE: To disable collection of a specific resource type, simply exclude all
|
|
metrics using the XX_metric_exclude. For example, to disable collection of VMs,
|
|
add this:
|
|
|
|
```toml
|
|
vm_metric_exclude = [ "*" ]
|
|
```
|
|
|
|
NOTE: To disable collection of a specific resource type, simply exclude all
|
|
metrics using the XX_metric_exclude.
|
|
For example, to disable collection of VMs, add this:
|
|
|
|
### Objects and Metrics per Query
|
|
|
|
By default, in the vCenter configuration a limit is set to the number of
|
|
entities that are included in a performance chart query. Default settings for
|
|
vCenter 6.5 and later is 256. Earlier versions of vCenter have this set to 64.
|
|
A vCenter administrator can change this setting.
|
|
See this [VMware KB article](https://kb.vmware.com/s/article/2107096) for more
|
|
information.
|
|
|
|
Any modification should be reflected in this plugin by modifying the parameter
|
|
`max_query_objects`
|
|
|
|
```toml
|
|
## number of objects to retrieve per query for realtime resources (VMs and hosts)
|
|
## set to 64 for vCenter 5.5 and 6.0 (default: 256)
|
|
# max_query_objects = 256
|
|
```
|
|
|
|
### Collection and Discovery Concurrency
|
|
|
|
In large vCenter setups it may be prudent to have multiple concurrent go
|
|
routines collect performance metrics in order to avoid potential errors for
|
|
time elapsed during a collection cycle. This should never be greater than 8,
|
|
though the default of 1 (no concurrency) should be sufficient for most
|
|
configurations.
|
|
|
|
For setting up concurrency, modify `collect_concurrency` and
|
|
`discover_concurrency` parameters.
|
|
|
|
```toml
|
|
## number of go routines to use for collection and discovery of objects and metrics
|
|
# collect_concurrency = 1
|
|
# discover_concurrency = 1
|
|
```
|
|
|
|
### Inventory Paths
|
|
|
|
Resources to be monitored can be selected using Inventory Paths. This treats
|
|
the vSphere inventory as a tree structure similar to a file system. A vSphere
|
|
inventory has a structure similar to this:
|
|
|
|
```bash
|
|
<root>
|
|
+-DC0 # Virtual datacenter
|
|
+-datastore # Datastore folder (created by system)
|
|
| +-Datastore1
|
|
+-host # Host folder (created by system)
|
|
| +-Cluster1
|
|
| | +-Host1
|
|
| | | +-VM1
|
|
| | | +-VM2
|
|
| | | +-hadoop1
|
|
| | +-ResourcePool1
|
|
| | | +-VM3
|
|
| | | +-VM4
|
|
| +-Host2 # Dummy cluster created for non-clustered host
|
|
| | +-Host2
|
|
| | | +-VM5
|
|
| | | +-VM6
|
|
+-vm # VM folder (created by system)
|
|
| +-VM1
|
|
| +-VM2
|
|
| +-Folder1
|
|
| | +-hadoop1
|
|
| | +-NestedFolder1
|
|
| | | +-VM3
|
|
| | | +-VM4
|
|
```
|
|
|
|
#### Using Inventory Paths
|
|
|
|
Using familiar UNIX-style paths, one could select e.g. VM2 with the path
|
|
`/DC0/vm/VM2`.
|
|
|
|
Often, we want to select a group of resource, such as all the VMs in a
|
|
folder. We could use the path `/DC0/vm/Folder1/*` for that.
|
|
|
|
Another possibility is to select objects using a partial name, such as
|
|
`/DC0/vm/Folder1/hadoop*` yielding all VMs in Folder1 with a name starting
|
|
with "hadoop".
|
|
|
|
Finally, due to the arbitrary nesting of the folder structure, we need a
|
|
"recursive wildcard" for traversing multiple folders. We use the "**" symbol
|
|
for that. If we want to look for a VM with a name starting with "hadoop" in
|
|
any folder, we could use the following path: `/DC0/vm/**/hadoop*`
|
|
|
|
#### Multiple Paths to VMs
|
|
|
|
As we can see from the example tree above, VMs appear both in its on folder
|
|
under the datacenter, as well as under the hosts. This is useful when you like
|
|
to select VMs on a specific host. For example,
|
|
`/DC0/host/Cluster1/Host1/hadoop*` selects all VMs with a name starting with
|
|
"hadoop" that are running on Host1.
|
|
|
|
We can extend this to looking at a cluster level:
|
|
`/DC0/host/Cluster1/*/hadoop*`. This selects any VM matching "hadoop*" on any
|
|
host in Cluster1.
|
|
|
|
#### Inventory paths and top-level folders
|
|
|
|
If your datacenter is in a folder and not directly below the inventory root, the
|
|
default inventory paths will not work. This is intentional, since recursive
|
|
wildcards may be slow in very large environments.
|
|
|
|
If your datacenter is in a folder, you have two options:
|
|
|
|
1. Explicitly include the folder in the path. For example, if your datacenter is in
|
|
a folder named ```F1``` you could use the following path to get to your hosts:
|
|
```/F1/MyDatacenter/host/**```
|
|
2. Use a recursive wildcard to search an arbitrarily long chain of nested folders. To
|
|
get to the hosts, you could use the following path: ```/**/host/**```. Note that
|
|
this may run slowly in a very large environment, since a large number of nodes will
|
|
be traversed.
|
|
|
|
## Performance Considerations
|
|
|
|
### Realtime vs. Historical Metrics
|
|
|
|
vCenter keeps two different kinds of metrics, known as realtime and historical
|
|
metrics.
|
|
|
|
* Realtime metrics: Available at a 20 second granularity. These metrics are stored in memory and are very fast and cheap to query. Our tests have shown that a complete set of realtime metrics for 7000 virtual machines can be obtained in less than 20 seconds. Realtime metrics are only available on **ESXi hosts** and **virtual machine** resources. Realtime metrics are only stored for 1 hour in vCenter.
|
|
* Historical metrics: Available at a (default) 5 minute, 30 minutes, 2 hours and 24 hours rollup levels. The vSphere Telegraf plugin only uses the most granular rollup which defaults to 5 minutes but can be changed in vCenter to other interval durations. These metrics are stored in the vCenter database and can be expensive and slow to query. Historical metrics are the only type of metrics available for **clusters**, **datastores**, **resource pools** and **datacenters**.
|
|
|
|
This distinction has an impact on how Telegraf collects metrics. A single
|
|
instance of an input plugin can have one and only one collection interval,
|
|
which means that you typically set the collection interval based on the most
|
|
frequently collected metric. Let's assume you set the collection interval to 1
|
|
minute. All realtime metrics will be collected every minute. Since the
|
|
historical metrics are only available on a 5 minute interval, the vSphere
|
|
Telegraf plugin automatically skips four out of five collection cycles for
|
|
these metrics. This works fine in many cases. Problems arise when the
|
|
collection of historical metrics takes longer than the collection interval.
|
|
This will cause error messages similar to this to appear in the Telegraf logs:
|
|
|
|
```text
|
|
2019-01-16T13:41:10Z W! [agent] input "inputs.vsphere" did not complete within its interval
|
|
```
|
|
|
|
This will disrupt the metric collection and can result in missed samples. The
|
|
best practice workaround is to specify two instances of the vSphere plugin, one
|
|
for the realtime metrics with a short collection interval and one for the
|
|
historical metrics with a longer interval. You can use the `*_metric_exclude`
|
|
to turn off the resources you don't want to collect metrics for in each
|
|
instance. For example:
|
|
|
|
```toml
|
|
## Realtime instance
|
|
[[inputs.vsphere]]
|
|
interval = "60s"
|
|
vcenters = [ "https://someaddress/sdk" ]
|
|
username = "someuser@vsphere.local"
|
|
password = "secret"
|
|
|
|
insecure_skip_verify = true
|
|
force_discover_on_init = true
|
|
|
|
# Exclude all historical metrics
|
|
datastore_metric_exclude = ["*"]
|
|
cluster_metric_exclude = ["*"]
|
|
datacenter_metric_exclude = ["*"]
|
|
resource_pool_metric_exclude = ["*"]
|
|
vsan_metric_exclude = ["*"]
|
|
|
|
collect_concurrency = 5
|
|
discover_concurrency = 5
|
|
|
|
# Historical instance
|
|
[[inputs.vsphere]]
|
|
|
|
interval = "300s"
|
|
|
|
vcenters = [ "https://someaddress/sdk" ]
|
|
username = "someuser@vsphere.local"
|
|
password = "secret"
|
|
|
|
insecure_skip_verify = true
|
|
force_discover_on_init = true
|
|
host_metric_exclude = ["*"] # Exclude realtime metrics
|
|
vm_metric_exclude = ["*"] # Exclude realtime metrics
|
|
|
|
max_query_metrics = 256
|
|
collect_concurrency = 3
|
|
```
|
|
|
|
### Configuring max_query_metrics Setting
|
|
|
|
The `max_query_metrics` determines the maximum number of metrics to attempt to
|
|
retrieve in one call to vCenter. Generally speaking, a higher number means
|
|
faster and more efficient queries. However, the number of allowed metrics in a
|
|
query is typically limited in vCenter by the `config.vpxd.stats.maxQueryMetrics`
|
|
setting in vCenter. The value defaults to 64 on vSphere 5.5 and earlier and to
|
|
256 on more recent versions. The vSphere plugin always checks this setting and
|
|
will automatically reduce the number if the limit configured in vCenter is lower
|
|
than max_query_metrics in the plugin. This will result in a log message similar
|
|
to this:
|
|
|
|
```text
|
|
2019-01-21T03:24:18Z W! [input.vsphere] Configured max_query_metrics is 256, but server limits it to 64. Reducing.
|
|
```
|
|
|
|
You may ask a vCenter administrator to increase this limit to help boost
|
|
performance.
|
|
|
|
### Cluster Metrics and the max_query_metrics Setting
|
|
|
|
Cluster metrics are handled a bit differently by vCenter. They are aggregated
|
|
from ESXi and virtual machine metrics and may not be available when you query
|
|
their most recent values. When this happens, vCenter will attempt to perform
|
|
that aggregation on the fly. Unfortunately, all the subqueries needed
|
|
internally in vCenter to perform this aggregation will count towards
|
|
`config.vpxd.stats.maxQueryMetrics`. This means that even a very small query
|
|
may result in an error message similar to this:
|
|
|
|
```text
|
|
2018-11-02T13:37:11Z E! Error in plugin [inputs.vsphere]: ServerFaultCode: This operation is restricted by the administrator - 'vpxd.stats.maxQueryMetrics'. Contact your system administrator
|
|
```
|
|
|
|
There are two ways of addressing this:
|
|
|
|
* Ask your vCenter administrator to set `config.vpxd.stats.maxQueryMetrics` to a number that's higher than the total number of virtual machines managed by a vCenter instance.
|
|
* Exclude the cluster metrics and use either the basicstats aggregator to calculate sums and averages per cluster or use queries in the visualization tool to obtain the same result.
|
|
|
|
### Concurrency Settings
|
|
|
|
The vSphere plugin allows you to specify two concurrency settings:
|
|
|
|
* `collect_concurrency`: The maximum number of simultaneous queries for performance metrics allowed per resource type.
|
|
* `discover_concurrency`: The maximum number of simultaneous queries for resource discovery allowed.
|
|
|
|
While a higher level of concurrency typically has a positive impact on
|
|
performance, increasing these numbers too much can cause performance issues at
|
|
the vCenter server. A rule of thumb is to set these parameters to the number of
|
|
virtual machines divided by 1500 and rounded up to the nearest integer.
|
|
|
|
### Configuring historical_interval Setting
|
|
|
|
When the vSphere plugin queries vCenter for historical statistics it queries for
|
|
statistics that exist at a specific interval. The default historical interval
|
|
duration is 5 minutes but if this interval has been changed then you must
|
|
override the default query interval in the vSphere plugin.
|
|
|
|
* `historical_interval`: The interval of the most granular statistics configured in vSphere represented in seconds.
|
|
|
|
## Metrics
|
|
|
|
* Cluster Stats
|
|
* Cluster services: CPU, memory, failover
|
|
* CPU: total, usage
|
|
* Memory: consumed, total, vmmemctl
|
|
* VM operations: # changes, clone, create, deploy, destroy, power, reboot, reconfigure, register, reset, shutdown, standby, vmotion
|
|
* Host Stats:
|
|
* CPU: total, usage, cost, mhz
|
|
* Datastore: iops, latency, read/write bytes, # reads/writes
|
|
* Disk: commands, latency, kernel reads/writes, # reads/writes, queues
|
|
* Memory: total, usage, active, latency, swap, shared, vmmemctl
|
|
* Network: broadcast, bytes, dropped, errors, multicast, packets, usage
|
|
* Power: energy, usage, capacity
|
|
* Res CPU: active, max, running
|
|
* Storage Adapter: commands, latency, # reads/writes
|
|
* Storage Path: commands, latency, # reads/writes
|
|
* System Resources: cpu active, cpu max, cpu running, cpu usage, mem allocated, mem consumed, mem shared, swap
|
|
* System: uptime
|
|
* Flash Module: active VMDKs
|
|
* VM Stats:
|
|
* CPU: demand, usage, readiness, cost, mhz
|
|
* Datastore: latency, # reads/writes
|
|
* Disk: commands, latency, # reads/writes, provisioned, usage
|
|
* Memory: granted, usage, active, swap, vmmemctl
|
|
* Network: broadcast, bytes, dropped, multicast, packets, usage
|
|
* Power: energy, usage
|
|
* Res CPU: active, max, running
|
|
* System: operating system uptime, uptime
|
|
* Virtual Disk: seeks, # reads/writes, latency, load
|
|
* Resource Pools stats:
|
|
* Memory: total, usage, active, latency, swap, shared, vmmemctl
|
|
* CPU: capacity, usage, corecount
|
|
* Disk: throughput
|
|
* Network: throughput
|
|
* Power: energy, usage
|
|
* Datastore stats:
|
|
* Disk: Capacity, provisioned, used
|
|
|
|
For a detailed list of commonly available metrics, please refer to
|
|
[METRICS.md](METRICS.md)
|
|
|
|
### Tags
|
|
|
|
* all metrics
|
|
* vcenter (vcenter url)
|
|
* all host metrics
|
|
* cluster (vcenter cluster)
|
|
* all vm metrics
|
|
* cluster (vcenter cluster)
|
|
* esxhost (name of ESXi host)
|
|
* guest (guest operating system id)
|
|
* resource pool (name of resource pool)
|
|
* cpu stats for Host and VM
|
|
* cpu (cpu core - not all CPU fields will have this tag)
|
|
* datastore stats for Host and VM
|
|
* datastore (id of datastore)
|
|
* disk stats for Host and VM
|
|
* disk (name of disk)
|
|
* disk.used.capacity for Datastore
|
|
* disk (type of disk)
|
|
* net stats for Host and VM
|
|
* interface (name of network interface)
|
|
* storageAdapter stats for Host
|
|
* adapter (name of storage adapter)
|
|
* storagePath stats for Host
|
|
* path (id of storage path)
|
|
* sys.resource* stats for Host
|
|
* resource (resource type)
|
|
* vflashModule stats for Host
|
|
* module (name of flash module)
|
|
* virtualDisk stats for VM
|
|
* disk (name of virtual disk)
|
|
|
|
## Add a vSAN extension
|
|
|
|
A vSAN resource is a special type of resource that can be collected by the
|
|
plugin. The configuration of a vSAN resource slightly differs from the
|
|
configuration of hosts, VMs, and other resources.
|
|
|
|
### Prerequisites for vSAN
|
|
|
|
* vSphere 6.5 and later
|
|
* Clusters with vSAN enabled
|
|
* [Turn on Virtual SAN performance service](https://docs.vmware.com/en/VMware-vSphere/6.5/com.vmware.vsphere.virtualsan.doc/GUID-02F67DC3-3D5A-48A4-A445-D2BD6AF2862C.html): When you create a vSAN cluster,
|
|
the performance service is disabled. To monitor the performance metrics,
|
|
you must turn on vSAN performance service.
|
|
|
|
### vSAN Configuration
|
|
|
|
```toml
|
|
[[inputs.vsphere]]
|
|
interval = "300s"
|
|
vcenters = ["https://<vcenter-ip>/sdk", "https://<vcenter2-ip>/sdk"]
|
|
username = "<user>"
|
|
password = "<pwd>"
|
|
|
|
# Exclude all other metrics
|
|
vm_metric_exclude = ["*"]
|
|
datastore_metric_exclude = ["*"]
|
|
datacenter_metric_exclude = ["*"]
|
|
host_metric_exclude = ["*"]
|
|
cluster_metric_exclude = ["*"]
|
|
|
|
# By default all supported entity will be included
|
|
vsan_metric_include = [
|
|
"summary.disk-usage",
|
|
"summary.health",
|
|
"summary.resync",
|
|
"performance.cluster-domclient",
|
|
"performance.cluster-domcompmgr",
|
|
"performance.host-domclient",
|
|
"performance.host-domcompmgr",
|
|
"performance.cache-disk",
|
|
"performance.disk-group",
|
|
"performance.capacity-disk",
|
|
"performance.disk-group",
|
|
"performance.virtual-machine",
|
|
"performance.vscsi",
|
|
"performance.virtual-disk",
|
|
"performance.vsan-host-net",
|
|
"performance.vsan-vnic-net",
|
|
"performance.vsan-pnic-net",
|
|
"performance.vsan-iscsi-host",
|
|
"performance.vsan-iscsi-target",
|
|
"performance.vsan-iscsi-lun",
|
|
"performance.lsom-world-cpu",
|
|
"performance.nic-world-cpu",
|
|
"performance.dom-world-cpu",
|
|
"performance.cmmds-world-cpu",
|
|
"performance.host-cpu",
|
|
"performance.host-domowner",
|
|
"performance.host-memory-slab",
|
|
"performance.host-memory-heap",
|
|
"performance.system-mem",
|
|
]
|
|
# by default vsan_metric_skip_verify = false
|
|
vsan_metric_skip_verify = true
|
|
vsan_metric_exclude = [ ]
|
|
# vsan_cluster_include = [ "/*/host/**" ] # Inventory path to clusters to collect (by default all are collected)
|
|
|
|
collect_concurrency = 5
|
|
discover_concurrency = 5
|
|
|
|
## Optional SSL Config
|
|
# ssl_ca = "/path/to/cafile"
|
|
# ssl_cert = "/path/to/certfile"
|
|
# ssl_key = "/path/to/keyfile"
|
|
## Use SSL but skip chain & host verification
|
|
# insecure_skip_verify = false
|
|
```
|
|
|
|
* Use `vsan_metric_include = [...]` to define the vSAN metrics that you want to collect.
|
|
For example, `vsan_metric_include = ["summary.*", "performance.host-domclient", "performance.cache-disk", "performance.disk-group", "performance.capacity-disk"]`.
|
|
To include all supported vSAN metrics, use `vsan_metric_include = [ "*" ]`.
|
|
To disable all the vSAN metrics, use `vsan_metric_exclude = [ "*" ]`.
|
|
|
|
* `vsan_metric_skip_verify` defines whether to skip verifying vSAN metrics against the ones from [GetSupportedEntityTypes API](https://code.vmware.com/apis/48/vsan#/doc/vim.cluster.VsanPerformanceManager.html#getSupportedEntityTypes).
|
|
This option is given because some performance entities are not returned by the API, but we want to offer the flexibility if you really need the stats.
|
|
When set to false, anything not in the supported entity list will be filtered out.
|
|
When set to true, queried metrics will be identical to vsan_metric_include and the exclusive array will not be used in this case. By default the value is false.
|
|
|
|
* `vsan_cluster_include` defines a list of inventory paths that will be used to select a portion of vSAN clusters.
|
|
vSAN metrics are only collected on the cluster level. Therefore, use the same way as inventory paths for [vSphere clusters](README.md#inventory-paths).
|
|
|
|
* Many vCenter environments use self-signed certificates. Update the bottom portion of the above configuration and provide proper values for all applicable SSL Config settings that apply in your vSphere environment. In some environments, setting insecure_skip_verify = true will be necessary when the SSL certificates are not available.
|
|
|
|
* To ensure consistent collection in larger vSphere environments, you must increase concurrency for the plugin. Use the collect_concurrency setting to control concurrency. Set collect_concurrency to the number of virtual machines divided by 1500 and rounded up to the nearest integer. For example, for 1200 VMs use 1, and for 2300 VMs use 2.
|
|
|
|
### Measurements & Fields
|
|
|
|
**NOTE**: Depending on the vSAN version, the vSAN performance measurements
|
|
and fields may vary.
|
|
|
|
* vSAN Summary
|
|
* overall_health
|
|
* total_capacity_bytes, free_capacity_bytes
|
|
* total_bytes_to_sync, total_objects_to_sync, total_recovery_eta
|
|
|
|
* vSAN Performance
|
|
* cluster-domclient
|
|
* iops_read, throughput_read, latency_avg_read, iops_write, throughput_write, latency_avg_write, congestion, oio
|
|
* cluster-domcompmgr
|
|
* iops_read, throughput_read, latency_avg_read, iops_write, throughput_write, latency_avg_write, iops_rec_write, throughput_rec_write, latency_avg_rec_write, congestion, oio, iops_resync_read, tput_resync_read, lat_avg_resyncread
|
|
* host-domclient
|
|
* iops_read, throughput_read, latency_avg_read, read_count, iops_write, throughput_write, latency_avg_write, write_count, congestion, oio, client_cache_hits, client_cache_hit_rate
|
|
* host-domcompmgr
|
|
* iops_read, throughput_read, latency_avg_read, read_count, iops_write, throughput_write, latency_avg_write, write_count, iops_rec_write, throughput_rec_write, latency_avg_rec_write, rec_write_count congestion, oio, iops_resync_read, tput_resync_read, lat_avg_resync_read
|
|
* cache-disk
|
|
* iops_dev_read, throughput_dev_read, latency_dev_read, io_count_dev_read, iops_dev_write, throughput_dev_write, latency_dev_write, io_count_dev_write, latency_dev_d_avg, latency_dev_g_avg
|
|
* capacity-disk
|
|
* iops_dev_read, throughput_dev_read, latency_dev_read, io_count_dev_read, iops_dev_write, throughput_dev_write, latency_dev_write, io_count_dev_write, latency_dev_d_avg, latency_dev_g_avg, iops_read, latency_read, io_count_read, iops_write, latency_write, io_count_write
|
|
* disk-group
|
|
* iops_sched, latency_sched, outstanding_bytes_sched, iops_sched_queue_rec, throughput_sched_queue_rec,latency_sched_queue_rec, iops_sched_queue_vm, throughput_sched_queue_vm,latency_sched_queue_vm, iops_sched_queue_meta, throughput_sched_queue_meta,latency_sched_queue_meta, iops_delay_pct_sched, latency_delay_sched, rc_hit_rate, wb_free_pct, war_evictions, quota_evictions, iops_rc_read, latency_rc_read, io_count_rc_read, iops_wb_read, latency_wb_read, io_count_wb_read, iops_rc_write, latency_rc_write, io_count_rc_write, iops_wb_write, latency_wb_write, io_count_wb_write, ssd_bytes_drained, zero_bytes_drained, mem_congestion, slab_congestion, ssd_congestion, iops_congestion, log_congestion, comp_congestion, iops_direct_sched, iops_read, throughput_read, latency_avg_read, read_count, iops_write, throughput_write, latency_avg_write, write_count, oio_write, oio_rec_write, oio_write_size, oio_rec_write_size, rc_size, wb_size, capacity, capacity_used, capacity_reserved, throughput_sched, iops_resync_read_policy, iops_resync_read_decom, iops_resync_read_rebalance, iops_resync_read_fix_comp, iops_resync_write_policy, iops_resync_write_decom, iops_resync_write_rebalance, iops_resync_write_fix_comp, tput_resync_read_policy, tput_resync_read_decom, tput_resync_read_rebalance, tput_resync_read_fix_comp, tput_resync_write_policy, tput_resync_write_decom, tput_resync_write_rebalance, tput_resync_write_fix_comp, lat_resync_read_policy, lat_resync_read_decom, lat_resync_read_rebalance, lat_resync_read_fix_comp, lat_resync_write_policy, lat_resync_write_decom, lat_resync_write_rebalance, lat_resync_write_fix_comp
|
|
* virtual-machine
|
|
* iops_read, throughput_read, latency_read_avg, latency_read_stddev, read_count, iops_write, throughput_write, latency_write_avg, latency_write_stddev, write_count
|
|
* vscsi
|
|
* iops_read, throughput_read, latency_read, read_count, iops_write, throughput_write, latency_write, write_count
|
|
* virtual-disk
|
|
* iops_limit, niops, niops_delayed
|
|
* vsan-host-net
|
|
* rx_throughput, rx_packets, rx_packets_loss_rate, tx_throughput, tx_packets, tx_packets_loss_rate
|
|
* vsan-vnic-net
|
|
* rx_throughput, rx_packets, rx_packets_loss_rate, tx_throughput, tx_packets, tx_packets_loss_rate
|
|
* vsan-pnic-net
|
|
* rx_throughput, rx_packets, rx_packets_loss_rate, tx_throughput, tx_packets, tx_packets_loss_rate
|
|
* vsan-iscsi-host
|
|
* iops_read, iops_write, iops_total, bandwidth_read, bandwidth_write, bandwidth_total, latency_read, latency_write, latency_total, queue_depth
|
|
* vsan-iscsi-target
|
|
* iops_read, iops_write, iops_total, bandwidth_read, bandwidth_write, bandwidth_total, latency_read, latency_write, latency_total, queue_depth
|
|
* vsan-iscsi-lun
|
|
* iops_read, iops_write, iops_total, bandwidth_read, bandwidth_write, bandwidth_total, latency_read, latency_write, latency_total, queue_depth
|
|
|
|
### vSAN Tags
|
|
|
|
* all vSAN metrics
|
|
* vcenter
|
|
* dcname
|
|
* clustername
|
|
* moid (the cluster's managed object id)
|
|
* host-domclient, host-domcompmgr
|
|
* hostname
|
|
* disk-group, cache-disk, capacity-disk
|
|
* hostname
|
|
* deviceName
|
|
* ssdUuid (if SSD)
|
|
* vsan-host-net
|
|
* hostname
|
|
* vsan-pnic-net
|
|
* pnic
|
|
* vsan-vnic-net
|
|
* vnic
|
|
* stackName
|
|
|
|
### Realtime vs. Historical Metrics in vSAN
|
|
|
|
vSAN metrics also keep two different kinds of metrics - realtime and
|
|
historical metrics.
|
|
|
|
* Realtime metrics are metrics with the prefix 'summary'. These metrics are available in realtime.
|
|
* Historical metrics are metrics with the prefix 'performance'. These are metrics queried from vSAN performance API, which is available at a 5-minute rollup level.
|
|
|
|
For performance consideration, it is better to specify two instances of the
|
|
plugin, one for the realtime metrics with a short collection interval,
|
|
and the second one - for the historical metrics with a longer interval.
|
|
For example:
|
|
|
|
```toml
|
|
## Realtime instance
|
|
[[inputs.vsphere]]
|
|
interval = "30s"
|
|
vcenters = [ "https://someaddress/sdk" ]
|
|
username = "someuser@vsphere.local"
|
|
password = "secret"
|
|
|
|
insecure_skip_verify = true
|
|
force_discover_on_init = true
|
|
|
|
# Exclude all other metrics
|
|
vm_metric_exclude = ["*"]
|
|
datastore_metric_exclude = ["*"]
|
|
datacenter_metric_exclude = ["*"]
|
|
host_metric_exclude = ["*"]
|
|
cluster_metric_exclude = ["*"]
|
|
|
|
vsan_metric_include = [ "summary.*" ]
|
|
vsan_metric_exclude = [ ]
|
|
vsan_metric_skip_verify = false
|
|
|
|
collect_concurrency = 5
|
|
discover_concurrency = 5
|
|
|
|
# Historical instance
|
|
[[inputs.vsphere]]
|
|
|
|
interval = "300s"
|
|
vcenters = [ "https://someaddress/sdk" ]
|
|
username = "someuser@vsphere.local"
|
|
password = "secret"
|
|
|
|
insecure_skip_verify = true
|
|
force_discover_on_init = true
|
|
|
|
# Exclude all other metrics
|
|
vm_metric_exclude = ["*"]
|
|
datastore_metric_exclude = ["*"]
|
|
datacenter_metric_exclude = ["*"]
|
|
host_metric_exclude = ["*"]
|
|
cluster_metric_exclude = ["*"]
|
|
|
|
vsan_metric_include = [ "performance.*" ]
|
|
vsan_metric_exclude = [ ]
|
|
vsan_metric_skip_verify = false
|
|
|
|
collect_concurrency = 5
|
|
discover_concurrency = 5
|
|
```
|
|
|
|
## Example Output
|
|
|
|
```text
|
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 run_summation=2608i,ready_summation=129i,usage_average=5.01,used_summation=2134i,demand_average=326i 1535660299000000000
|
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 bytesRx_average=321i,bytesTx_average=335i 1535660299000000000
|
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 write_average=144i,read_average=4i 1535660299000000000
|
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 bytesRx_average=242i,bytesTx_average=308i 1535660299000000000
|
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 write_average=232i,read_average=4i 1535660299000000000
|
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 usage_average=5.49,used_summation=1804i,demand_average=308i,run_summation=2001i,ready_summation=120i 1535660299000000000
|
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 usage_average=4.19,used_summation=2108i,demand_average=285i,run_summation=1793i,ready_summation=93i 1535660299000000000
|
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 bytesRx_average=272i,bytesTx_average=419i 1535660299000000000
|
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 write_average=229i,read_average=4i 1535660299000000000
|
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 run_summation=2277i,ready_summation=118i,usage_average=4.67,used_summation=2546i,demand_average=289i 1535660299000000000
|
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 bytesRx_average=243i,bytesTx_average=296i 1535660299000000000
|
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 write_average=158i,read_average=4i 1535660299000000000
|
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,interface=vmnic0,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=1042i,bytesTx_average=753i,bytesRx_average=660i 1535660299000000000
|
|
vsphere_host_cpu,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 utilization_average=10.46,usage_average=22.4,readiness_average=0.4,costop_summation=2i,coreUtilization_average=19.61,wait_summation=5148518i,idle_summation=58581i,latency_average=0.6,ready_summation=13370i,used_summation=19219i 1535660299000000000
|
|
vsphere_host_cpu,cpu=0,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=25.6,utilization_average=11.58,used_summation=24306i,usage_average=24.26,idle_summation=86688i 1535660299000000000
|
|
vsphere_host_cpu,cpu=1,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=12.29,utilization_average=8.32,used_summation=31312i,usage_average=22.47,idle_summation=94934i 1535660299000000000
|
|
vsphere_host_disk,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 read_average=331i,write_average=2800i 1535660299000000000
|
|
vsphere_host_disk,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 write_average=2701i,read_average=258i 1535660299000000000
|
|
vsphere_host_mem,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=93.27 1535660299000000000
|
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 bytesTx_average=650i,usage_average=1414i,bytesRx_average=569i 1535660299000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,cpu=1,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 utilization_average=12.6,used_summation=25775i,usage_average=24.44,idle_summation=68886i,coreUtilization_average=17.59 1535660299000000000
|
|
vsphere_host_disk,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 read_average=340i,write_average=2340i 1535660299000000000
|
|
vsphere_host_disk,clustername=DC0_C0,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 write_average=2277i,read_average=282i 1535660299000000000
|
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=104.78 1535660299000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesTx_average=463i,usage_average=1131i,bytesRx_average=719i 1535660299000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,interface=vmnic0,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=1668i,bytesTx_average=838i,bytesRx_average=921i 1535660299000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 used_summation=28952i,utilization_average=11.36,idle_summation=93261i,latency_average=0.46,ready_summation=12837i,usage_average=21.56,readiness_average=0.39,costop_summation=2i,coreUtilization_average=27.19,wait_summation=3820829i 1535660299000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,cpu=0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 coreUtilization_average=24.12,utilization_average=13.83,used_summation=22462i,usage_average=24.69,idle_summation=96993i 1535660299000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=localhost:8989 connect_ns=4727607i,discover_ns=65389011i,discovered_objects=8i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=localhost:8989 gather_duration_ns=296223i,gather_count=0i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=192.168.1.151 gather_duration_ns=136050i,gather_count=0i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=localhost:8989 gather_count=62i,gather_duration_ns=8788033i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=192.168.1.151 gather_count=0i,gather_duration_ns=162002i 1535660309000000000
|
|
internal_gather,host=host.example.com,input=vsphere,os=Mac gather_time_ns=17483653i,metrics_gathered=28i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=192.168.1.151 connect_ns=0i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=localhost:8989 gather_duration_ns=7291897i,gather_count=36i 1535660309000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=192.168.1.151 gather_duration_ns=958474i,gather_count=0i 1535660309000000000
|
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 usage_average=8.82,used_summation=3192i,demand_average=283i,run_summation=2419i,ready_summation=115i 1535660319000000000
|
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 bytesRx_average=277i,bytesTx_average=343i 1535660319000000000
|
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 read_average=1i,write_average=741i 1535660319000000000
|
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 bytesRx_average=386i,bytesTx_average=369i 1535660319000000000
|
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 write_average=814i,read_average=1i 1535660319000000000
|
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 run_summation=1778i,ready_summation=111i,usage_average=7.54,used_summation=2339i,demand_average=297i 1535660319000000000
|
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 usage_average=6.98,used_summation=2125i,demand_average=211i,run_summation=2990i,ready_summation=141i 1535660319000000000
|
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 bytesRx_average=357i,bytesTx_average=268i 1535660319000000000
|
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 write_average=528i,read_average=1i 1535660319000000000
|
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 used_summation=2374i,demand_average=195i,run_summation=3454i,ready_summation=110i,usage_average=7.34 1535660319000000000
|
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 bytesRx_average=308i,bytesTx_average=246i 1535660319000000000
|
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 write_average=1178i,read_average=1i 1535660319000000000
|
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,interface=vmnic0,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 bytesRx_average=773i,usage_average=1521i,bytesTx_average=890i 1535660319000000000
|
|
vsphere_host_cpu,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 wait_summation=3421258i,idle_summation=67994i,latency_average=0.36,usage_average=29.86,readiness_average=0.37,used_summation=25244i,costop_summation=2i,coreUtilization_average=21.94,utilization_average=17.19,ready_summation=15897i 1535660319000000000
|
|
vsphere_host_cpu,cpu=0,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 utilization_average=11.32,used_summation=19333i,usage_average=14.29,idle_summation=92708i,coreUtilization_average=27.68 1535660319000000000
|
|
vsphere_host_cpu,cpu=1,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 used_summation=28596i,usage_average=25.32,idle_summation=79553i,coreUtilization_average=28.01,utilization_average=11.33 1535660319000000000
|
|
vsphere_host_disk,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 read_average=86i,write_average=1659i 1535660319000000000
|
|
vsphere_host_disk,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 write_average=1997i,read_average=58i 1535660319000000000
|
|
vsphere_host_mem,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=68.45 1535660319000000000
|
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 bytesTx_average=679i,usage_average=2286i,bytesRx_average=719i 1535660319000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,cpu=1,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 utilization_average=10.52,used_summation=21693i,usage_average=23.09,idle_summation=84590i,coreUtilization_average=29.92 1535660319000000000
|
|
vsphere_host_disk,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 read_average=113i,write_average=1236i 1535660319000000000
|
|
vsphere_host_disk,clustername=DC0_C0,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 write_average=1708i,read_average=110i 1535660319000000000
|
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=111.46 1535660319000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesTx_average=998i,usage_average=2000i,bytesRx_average=881i 1535660319000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,interface=vmnic0,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=1683i,bytesTx_average=675i,bytesRx_average=1078i 1535660319000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 used_summation=28531i,wait_summation=3139129i,utilization_average=9.99,idle_summation=98579i,latency_average=0.51,costop_summation=2i,coreUtilization_average=14.35,ready_summation=16121i,usage_average=34.19,readiness_average=0.4 1535660319000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,cpu=0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 utilization_average=12.2,used_summation=22750i,usage_average=18.84,idle_summation=99539i,coreUtilization_average=23.05 1535660319000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=localhost:8989 gather_duration_ns=7076543i,gather_count=62i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=192.168.1.151 gather_duration_ns=4051303i,gather_count=0i 1535660339000000000
|
|
internal_gather,host=host.example.com,input=vsphere,os=Mac metrics_gathered=56i,gather_time_ns=13555029i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=192.168.1.151 connect_ns=0i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=localhost:8989 gather_duration_ns=6335467i,gather_count=36i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=192.168.1.151 gather_duration_ns=958474i,gather_count=0i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=localhost:8989 discover_ns=65389011i,discovered_objects=8i,connect_ns=4727607i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=localhost:8989 gather_duration_ns=296223i,gather_count=0i 1535660339000000000
|
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=192.168.1.151 gather_count=0i,gather_duration_ns=1540920i 1535660339000000000
|
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 write_average=302i,read_average=11i 1535660339000000000
|
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 usage_average=5.58,used_summation=2941i,demand_average=298i,run_summation=3255i,ready_summation=96i 1535660339000000000
|
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 bytesRx_average=155i,bytesTx_average=241i 1535660339000000000
|
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 usage_average=10.3,used_summation=3053i,demand_average=346i,run_summation=3289i,ready_summation=122i 1535660339000000000
|
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 bytesRx_average=215i,bytesTx_average=275i 1535660339000000000
|
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 write_average=252i,read_average=14i 1535660339000000000
|
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 usage_average=8,used_summation=2183i,demand_average=354i,run_summation=3542i,ready_summation=128i 1535660339000000000
|
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 bytesRx_average=178i,bytesTx_average=200i 1535660339000000000
|
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 write_average=283i,read_average=12i 1535660339000000000
|
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 demand_average=328i,run_summation=3481i,ready_summation=122i,usage_average=7.95,used_summation=2167i 1535660339000000000
|
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 bytesTx_average=282i,bytesRx_average=196i 1535660339000000000
|
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 write_average=321i,read_average=13i 1535660339000000000
|
|
vsphere_host_disk,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 read_average=39i,write_average=2635i 1535660339000000000
|
|
vsphere_host_disk,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 write_average=2635i,read_average=30i 1535660339000000000
|
|
vsphere_host_mem,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=98.5 1535660339000000000
|
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=1887i,bytesRx_average=662i,bytesTx_average=251i 1535660339000000000
|
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,interface=vmnic0,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=1481i,bytesTx_average=899i,bytesRx_average=992i 1535660339000000000
|
|
vsphere_host_cpu,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 used_summation=50405i,costop_summation=2i,utilization_average=17.32,latency_average=0.61,ready_summation=14843i,usage_average=27.94,coreUtilization_average=32.12,wait_summation=3058787i,idle_summation=56600i,readiness_average=0.36 1535660339000000000
|
|
vsphere_host_cpu,cpu=0,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=37.61,utilization_average=17.05,used_summation=38013i,usage_average=32.66,idle_summation=89575i 1535660339000000000
|
|
vsphere_host_cpu,cpu=1,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=25.92,utilization_average=18.72,used_summation=39790i,usage_average=40.42,idle_summation=69457i 1535660339000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,interface=vmnic0,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=1246i,bytesTx_average=673i,bytesRx_average=781i 1535660339000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 coreUtilization_average=33.8,idle_summation=77121i,ready_summation=15857i,readiness_average=0.39,used_summation=29554i,costop_summation=2i,wait_summation=4338417i,utilization_average=17.87,latency_average=0.44,usage_average=28.78 1535660339000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,cpu=0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 idle_summation=86610i,coreUtilization_average=34.36,utilization_average=19.03,used_summation=28766i,usage_average=23.72 1535660339000000000
|
|
vsphere_host_cpu,clustername=DC0_C0,cpu=1,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 coreUtilization_average=33.15,utilization_average=16.8,used_summation=44282i,usage_average=30.08,idle_summation=93490i 1535660339000000000
|
|
vsphere_host_disk,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 read_average=56i,write_average=1672i 1535660339000000000
|
|
vsphere_host_disk,clustername=DC0_C0,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 write_average=2110i,read_average=48i 1535660339000000000
|
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=116.21 1535660339000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesRx_average=726i,bytesTx_average=643i,usage_average=1504i 1535660339000000000
|
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=116.21 1535660339000000000
|
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesRx_average=726i,bytesTx_average=643i,usage_average=1504i 1535660339000000000
|
|
```
|
|
|
|
## vSAN Sample Output
|
|
|
|
```text
|
|
vsphere_vsan_performance_hostdomclient,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,hostname=DC0_C0_H0,moid=domain-c8,source=Example-VSAN,vcenter=localhost:8898 iops_read=7,write_congestion=0,unmap_congestion=0,read_count=2199,iops=8,latency_max_write=8964,latency_avg_unmap=0,latency_avg_write=1883,write_count=364,num_oio=12623,throughput=564127,client_cache_hits=0,latency_max_read=17821,latency_max_unmap=0,read_congestion=0,latency_avg=1154,congestion=0,throughput_read=554721,latency_avg_read=1033,throughput_write=9406,client_cache_hit_rate=0,iops_unmap=0,throughput_unmap=0,latency_stddev=1315,io_count=2563,oio=4,iops_write=1,unmap_count=0 1578955200000000000
|
|
vsphere_vsan_performance_clusterdomcompmgr,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,uuid=XXXXXXXX-XXXX-XXXX-XXXX-XXXXXXXXX,vcenter=localhost:8898 latency_avg_rec_write=0,latency_avg_write=9886,congestion=0,iops_resync_read=0,lat_avg_resync_read=0,iops_read=289,latency_avg_read=1184,throughput_write=50137368,iops_rec_write=0,throughput_rec_write=0,tput_resync_read=0,throughput_read=9043654,iops_write=1272,oio=97 1578954900000000000
|
|
vsphere_vsan_performance_clusterdomclient,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,uuid=XXXXXXXX-XXXX-XXXX-XXXX-XXXXXXXXX,vcenter=localhost:8898 latency_avg_write=1011,congestion=0,oio=26,iops_read=6,throughput_read=489093,latency_avg_read=1085,iops_write=43,throughput_write=435142 1578955200000000000
|
|
vsphere_vsan_summary,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,vcenter=localhost:8898 total_bytes_to_sync=0i,total_objects_to_sync=0i,total_recovery_eta=0i 1578955489000000000
|
|
vsphere_vsan_summary,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,vcenter=localhost:8898 overall_health=1i 1578955489000000000
|
|
vsphere_vsan_summary,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,vcenter=localhost:8898 free_capacity_byte=11022535578757i,total_capacity_byte=14102625779712i 1578955488000000000
|
|
```
|