1029 lines
60 KiB
Markdown
1029 lines
60 KiB
Markdown
|
# VMware vSphere Input Plugin
|
||
|
|
||
|
The VMware vSphere plugin uses the vSphere API to gather metrics from multiple
|
||
|
vCenter servers.
|
||
|
|
||
|
* Clusters
|
||
|
* Hosts
|
||
|
* Resource Pools
|
||
|
* VMs
|
||
|
* Datastores
|
||
|
* vSAN
|
||
|
|
||
|
## Supported versions of vSphere
|
||
|
|
||
|
This plugin supports vSphere version 6.5, 6.7, 7.0 and 8.0.
|
||
|
It may work with versions 5.1, 5.5 and 6.0, but neither are
|
||
|
officially supported.
|
||
|
|
||
|
Compatibility information is available from the govmomi project
|
||
|
[here](https://github.com/vmware/govmomi/tree/v0.26.0#compatibility)
|
||
|
|
||
|
## Global configuration options <!-- @/docs/includes/plugin_config.md -->
|
||
|
|
||
|
In addition to the plugin-specific configuration settings, plugins support
|
||
|
additional global and plugin configuration settings. These settings are used to
|
||
|
modify metrics, tags, and field or create aliases and configure ordering, etc.
|
||
|
See the [CONFIGURATION.md][CONFIGURATION.md] for more details.
|
||
|
|
||
|
[CONFIGURATION.md]: ../../../docs/CONFIGURATION.md#plugins
|
||
|
|
||
|
## Secret-store support
|
||
|
|
||
|
This plugin supports secrets from secret-stores for the `username` and
|
||
|
`password` option.
|
||
|
See the [secret-store documentation][SECRETSTORE] for more details on how
|
||
|
to use them.
|
||
|
|
||
|
[SECRETSTORE]: ../../../docs/CONFIGURATION.md#secret-store-secrets
|
||
|
|
||
|
## Configuration
|
||
|
|
||
|
```toml @sample.conf
|
||
|
# Read metrics from one or many vCenters
|
||
|
[[inputs.vsphere]]
|
||
|
## List of vCenter URLs to be monitored. These three lines must be uncommented
|
||
|
## and edited for the plugin to work.
|
||
|
vcenters = [ "https://vcenter.local/sdk" ]
|
||
|
username = "user@corp.local"
|
||
|
password = "secret"
|
||
|
|
||
|
## VMs
|
||
|
## Typical VM metrics (if omitted or empty, all metrics are collected)
|
||
|
# vm_include = [ "/*/vm/**"] # Inventory path to VMs to collect (by default all are collected)
|
||
|
# vm_exclude = [] # Inventory paths to exclude
|
||
|
vm_metric_include = [
|
||
|
"cpu.demand.average",
|
||
|
"cpu.idle.summation",
|
||
|
"cpu.latency.average",
|
||
|
"cpu.readiness.average",
|
||
|
"cpu.ready.summation",
|
||
|
"cpu.run.summation",
|
||
|
"cpu.usagemhz.average",
|
||
|
"cpu.used.summation",
|
||
|
"cpu.wait.summation",
|
||
|
"mem.active.average",
|
||
|
"mem.granted.average",
|
||
|
"mem.latency.average",
|
||
|
"mem.swapin.average",
|
||
|
"mem.swapinRate.average",
|
||
|
"mem.swapout.average",
|
||
|
"mem.swapoutRate.average",
|
||
|
"mem.usage.average",
|
||
|
"mem.vmmemctl.average",
|
||
|
"net.bytesRx.average",
|
||
|
"net.bytesTx.average",
|
||
|
"net.droppedRx.summation",
|
||
|
"net.droppedTx.summation",
|
||
|
"net.usage.average",
|
||
|
"power.power.average",
|
||
|
"virtualDisk.numberReadAveraged.average",
|
||
|
"virtualDisk.numberWriteAveraged.average",
|
||
|
"virtualDisk.read.average",
|
||
|
"virtualDisk.readOIO.latest",
|
||
|
"virtualDisk.throughput.usage.average",
|
||
|
"virtualDisk.totalReadLatency.average",
|
||
|
"virtualDisk.totalWriteLatency.average",
|
||
|
"virtualDisk.write.average",
|
||
|
"virtualDisk.writeOIO.latest",
|
||
|
"sys.uptime.latest",
|
||
|
]
|
||
|
# vm_metric_exclude = [] ## Nothing is excluded by default
|
||
|
# vm_instances = true ## true by default
|
||
|
|
||
|
## Hosts
|
||
|
## Typical host metrics (if omitted or empty, all metrics are collected)
|
||
|
# host_include = [ "/*/host/**"] # Inventory path to hosts to collect (by default all are collected)
|
||
|
# host_exclude [] # Inventory paths to exclude
|
||
|
host_metric_include = [
|
||
|
"cpu.coreUtilization.average",
|
||
|
"cpu.costop.summation",
|
||
|
"cpu.demand.average",
|
||
|
"cpu.idle.summation",
|
||
|
"cpu.latency.average",
|
||
|
"cpu.readiness.average",
|
||
|
"cpu.ready.summation",
|
||
|
"cpu.swapwait.summation",
|
||
|
"cpu.usage.average",
|
||
|
"cpu.usagemhz.average",
|
||
|
"cpu.used.summation",
|
||
|
"cpu.utilization.average",
|
||
|
"cpu.wait.summation",
|
||
|
"disk.deviceReadLatency.average",
|
||
|
"disk.deviceWriteLatency.average",
|
||
|
"disk.kernelReadLatency.average",
|
||
|
"disk.kernelWriteLatency.average",
|
||
|
"disk.numberReadAveraged.average",
|
||
|
"disk.numberWriteAveraged.average",
|
||
|
"disk.read.average",
|
||
|
"disk.totalReadLatency.average",
|
||
|
"disk.totalWriteLatency.average",
|
||
|
"disk.write.average",
|
||
|
"mem.active.average",
|
||
|
"mem.latency.average",
|
||
|
"mem.state.latest",
|
||
|
"mem.swapin.average",
|
||
|
"mem.swapinRate.average",
|
||
|
"mem.swapout.average",
|
||
|
"mem.swapoutRate.average",
|
||
|
"mem.totalCapacity.average",
|
||
|
"mem.usage.average",
|
||
|
"mem.vmmemctl.average",
|
||
|
"net.bytesRx.average",
|
||
|
"net.bytesTx.average",
|
||
|
"net.droppedRx.summation",
|
||
|
"net.droppedTx.summation",
|
||
|
"net.errorsRx.summation",
|
||
|
"net.errorsTx.summation",
|
||
|
"net.usage.average",
|
||
|
"power.power.average",
|
||
|
"storageAdapter.numberReadAveraged.average",
|
||
|
"storageAdapter.numberWriteAveraged.average",
|
||
|
"storageAdapter.read.average",
|
||
|
"storageAdapter.write.average",
|
||
|
"sys.uptime.latest",
|
||
|
]
|
||
|
## Collect IP addresses? Valid values are "ipv4" and "ipv6"
|
||
|
# ip_addresses = ["ipv6", "ipv4" ]
|
||
|
|
||
|
# host_metric_exclude = [] ## Nothing excluded by default
|
||
|
# host_instances = true ## true by default
|
||
|
|
||
|
|
||
|
## Clusters
|
||
|
# cluster_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
|
||
|
# cluster_exclude = [] # Inventory paths to exclude
|
||
|
# cluster_metric_include = [] ## if omitted or empty, all metrics are collected
|
||
|
# cluster_metric_exclude = [] ## Nothing excluded by default
|
||
|
# cluster_instances = false ## false by default
|
||
|
|
||
|
## Resource Pools
|
||
|
# resource_pool_include = [ "/*/host/**"] # Inventory path to resource pools to collect (by default all are collected)
|
||
|
# resource_pool_exclude = [] # Inventory paths to exclude
|
||
|
# resource_pool_metric_include = [] ## if omitted or empty, all metrics are collected
|
||
|
# resource_pool_metric_exclude = [] ## Nothing excluded by default
|
||
|
# resource_pool_instances = false ## false by default
|
||
|
|
||
|
## Datastores
|
||
|
# datastore_include = [ "/*/datastore/**"] # Inventory path to datastores to collect (by default all are collected)
|
||
|
# datastore_exclude = [] # Inventory paths to exclude
|
||
|
# datastore_metric_include = [] ## if omitted or empty, all metrics are collected
|
||
|
# datastore_metric_exclude = [] ## Nothing excluded by default
|
||
|
# datastore_instances = false ## false by default
|
||
|
|
||
|
## Datacenters
|
||
|
# datacenter_include = [ "/*/host/**"] # Inventory path to clusters to collect (by default all are collected)
|
||
|
# datacenter_exclude = [] # Inventory paths to exclude
|
||
|
datacenter_metric_include = [] ## if omitted or empty, all metrics are collected
|
||
|
datacenter_metric_exclude = [ "*" ] ## Datacenters are not collected by default.
|
||
|
# datacenter_instances = false ## false by default
|
||
|
|
||
|
## VSAN
|
||
|
# vsan_metric_include = [] ## if omitted or empty, all metrics are collected
|
||
|
# vsan_metric_exclude = [ "*" ] ## vSAN are not collected by default.
|
||
|
## Whether to skip verifying vSAN metrics against the ones from GetSupportedEntityTypes API.
|
||
|
# vsan_metric_skip_verify = false ## false by default.
|
||
|
|
||
|
## Interval for sampling vSAN performance metrics, can be reduced down to
|
||
|
## 30 seconds for vSAN 8 U1.
|
||
|
# vsan_interval = "5m"
|
||
|
|
||
|
## Plugin Settings
|
||
|
## separator character to use for measurement and field names (default: "_")
|
||
|
# separator = "_"
|
||
|
|
||
|
## number of objects to retrieve per query for realtime resources (vms and hosts)
|
||
|
## set to 64 for vCenter 5.5 and 6.0 (default: 256)
|
||
|
# max_query_objects = 256
|
||
|
|
||
|
## number of metrics to retrieve per query for non-realtime resources (clusters and datastores)
|
||
|
## set to 64 for vCenter 5.5 and 6.0 (default: 256)
|
||
|
# max_query_metrics = 256
|
||
|
|
||
|
## number of go routines to use for collection and discovery of objects and metrics
|
||
|
# collect_concurrency = 1
|
||
|
# discover_concurrency = 1
|
||
|
|
||
|
## the interval before (re)discovering objects subject to metrics collection (default: 300s)
|
||
|
# object_discovery_interval = "300s"
|
||
|
|
||
|
## timeout applies to any of the api request made to vcenter
|
||
|
# timeout = "60s"
|
||
|
|
||
|
## When set to true, all samples are sent as integers. This makes the output
|
||
|
## data types backwards compatible with Telegraf 1.9 or lower. Normally all
|
||
|
## samples from vCenter, with the exception of percentages, are integer
|
||
|
## values, but under some conditions, some averaging takes place internally in
|
||
|
## the plugin. Setting this flag to "false" will send values as floats to
|
||
|
## preserve the full precision when averaging takes place.
|
||
|
# use_int_samples = true
|
||
|
|
||
|
## Custom attributes from vCenter can be very useful for queries in order to slice the
|
||
|
## metrics along different dimension and for forming ad-hoc relationships. They are disabled
|
||
|
## by default, since they can add a considerable amount of tags to the resulting metrics. To
|
||
|
## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
|
||
|
## to select the attributes you want to include.
|
||
|
## By default, since they can add a considerable amount of tags to the resulting metrics. To
|
||
|
## enable, simply set custom_attribute_exclude to [] (empty set) and use custom_attribute_include
|
||
|
## to select the attributes you want to include.
|
||
|
# custom_attribute_include = []
|
||
|
# custom_attribute_exclude = ["*"]
|
||
|
|
||
|
## The number of vSphere 5 minute metric collection cycles to look back for non-realtime metrics. In
|
||
|
## some versions (6.7, 7.0 and possible more), certain metrics, such as cluster metrics, may be reported
|
||
|
## with a significant delay (>30min). If this happens, try increasing this number. Please note that increasing
|
||
|
## it too much may cause performance issues.
|
||
|
# metric_lookback = 3
|
||
|
|
||
|
## Optional SSL Config
|
||
|
# ssl_ca = "/path/to/cafile"
|
||
|
# ssl_cert = "/path/to/certfile"
|
||
|
# ssl_key = "/path/to/keyfile"
|
||
|
## Use SSL but skip chain & host verification
|
||
|
# insecure_skip_verify = false
|
||
|
|
||
|
## The Historical Interval value must match EXACTLY the interval in the daily
|
||
|
# "Interval Duration" found on the VCenter server under Configure > General > Statistics > Statistic intervals
|
||
|
# historical_interval = "5m"
|
||
|
|
||
|
## Specifies plugin behavior regarding disconnected servers
|
||
|
## Available choices :
|
||
|
## - error: telegraf will return an error on startup if one the servers is unreachable
|
||
|
## - ignore: telegraf will ignore unreachable servers on both startup and gather
|
||
|
# disconnected_servers_behavior = "error"
|
||
|
|
||
|
## HTTP Proxy support
|
||
|
# use_system_proxy = true
|
||
|
# http_proxy_url = ""
|
||
|
```
|
||
|
|
||
|
NOTE: To disable collection of a specific resource type, simply exclude all
|
||
|
metrics using the XX_metric_exclude. For example, to disable collection of VMs,
|
||
|
add this:
|
||
|
|
||
|
```toml
|
||
|
vm_metric_exclude = [ "*" ]
|
||
|
```
|
||
|
|
||
|
NOTE: To disable collection of a specific resource type, simply exclude all
|
||
|
metrics using the XX_metric_exclude.
|
||
|
For example, to disable collection of VMs, add this:
|
||
|
|
||
|
### Objects and Metrics per Query
|
||
|
|
||
|
By default, in the vCenter configuration a limit is set to the number of
|
||
|
entities that are included in a performance chart query. Default settings for
|
||
|
vCenter 6.5 and later is 256. Earlier versions of vCenter have this set to 64.
|
||
|
A vCenter administrator can change this setting.
|
||
|
See this [VMware KB article](https://kb.vmware.com/s/article/2107096) for more
|
||
|
information.
|
||
|
|
||
|
Any modification should be reflected in this plugin by modifying the parameter
|
||
|
`max_query_objects`
|
||
|
|
||
|
```toml
|
||
|
## number of objects to retrieve per query for realtime resources (VMs and hosts)
|
||
|
## set to 64 for vCenter 5.5 and 6.0 (default: 256)
|
||
|
# max_query_objects = 256
|
||
|
```
|
||
|
|
||
|
### Collection and Discovery Concurrency
|
||
|
|
||
|
In large vCenter setups it may be prudent to have multiple concurrent go
|
||
|
routines collect performance metrics in order to avoid potential errors for
|
||
|
time elapsed during a collection cycle. This should never be greater than 8,
|
||
|
though the default of 1 (no concurrency) should be sufficient for most
|
||
|
configurations.
|
||
|
|
||
|
For setting up concurrency, modify `collect_concurrency` and
|
||
|
`discover_concurrency` parameters.
|
||
|
|
||
|
```toml
|
||
|
## number of go routines to use for collection and discovery of objects and metrics
|
||
|
# collect_concurrency = 1
|
||
|
# discover_concurrency = 1
|
||
|
```
|
||
|
|
||
|
### Inventory Paths
|
||
|
|
||
|
Resources to be monitored can be selected using Inventory Paths. This treats
|
||
|
the vSphere inventory as a tree structure similar to a file system. A vSphere
|
||
|
inventory has a structure similar to this:
|
||
|
|
||
|
```bash
|
||
|
<root>
|
||
|
+-DC0 # Virtual datacenter
|
||
|
+-datastore # Datastore folder (created by system)
|
||
|
| +-Datastore1
|
||
|
+-host # Host folder (created by system)
|
||
|
| +-Cluster1
|
||
|
| | +-Host1
|
||
|
| | | +-VM1
|
||
|
| | | +-VM2
|
||
|
| | | +-hadoop1
|
||
|
| | +-ResourcePool1
|
||
|
| | | +-VM3
|
||
|
| | | +-VM4
|
||
|
| +-Host2 # Dummy cluster created for non-clustered host
|
||
|
| | +-Host2
|
||
|
| | | +-VM5
|
||
|
| | | +-VM6
|
||
|
+-vm # VM folder (created by system)
|
||
|
| +-VM1
|
||
|
| +-VM2
|
||
|
| +-Folder1
|
||
|
| | +-hadoop1
|
||
|
| | +-NestedFolder1
|
||
|
| | | +-VM3
|
||
|
| | | +-VM4
|
||
|
```
|
||
|
|
||
|
#### Using Inventory Paths
|
||
|
|
||
|
Using familiar UNIX-style paths, one could select e.g. VM2 with the path
|
||
|
`/DC0/vm/VM2`.
|
||
|
|
||
|
Often, we want to select a group of resource, such as all the VMs in a
|
||
|
folder. We could use the path `/DC0/vm/Folder1/*` for that.
|
||
|
|
||
|
Another possibility is to select objects using a partial name, such as
|
||
|
`/DC0/vm/Folder1/hadoop*` yielding all VMs in Folder1 with a name starting
|
||
|
with "hadoop".
|
||
|
|
||
|
Finally, due to the arbitrary nesting of the folder structure, we need a
|
||
|
"recursive wildcard" for traversing multiple folders. We use the "**" symbol
|
||
|
for that. If we want to look for a VM with a name starting with "hadoop" in
|
||
|
any folder, we could use the following path: `/DC0/vm/**/hadoop*`
|
||
|
|
||
|
#### Multiple Paths to VMs
|
||
|
|
||
|
As we can see from the example tree above, VMs appear both in its on folder
|
||
|
under the datacenter, as well as under the hosts. This is useful when you like
|
||
|
to select VMs on a specific host. For example,
|
||
|
`/DC0/host/Cluster1/Host1/hadoop*` selects all VMs with a name starting with
|
||
|
"hadoop" that are running on Host1.
|
||
|
|
||
|
We can extend this to looking at a cluster level:
|
||
|
`/DC0/host/Cluster1/*/hadoop*`. This selects any VM matching "hadoop*" on any
|
||
|
host in Cluster1.
|
||
|
|
||
|
#### Inventory paths and top-level folders
|
||
|
|
||
|
If your datacenter is in a folder and not directly below the inventory root, the
|
||
|
default inventory paths will not work. This is intentional, since recursive
|
||
|
wildcards may be slow in very large environments.
|
||
|
|
||
|
If your datacenter is in a folder, you have two options:
|
||
|
|
||
|
1. Explicitly include the folder in the path. For example, if your datacenter is
|
||
|
in a folder named ```F1``` you could use the path `/F1/MyDatacenter/host/**`
|
||
|
to get to your hosts.
|
||
|
2. Use a recursive wildcard to search an arbitrarily long chain of nested
|
||
|
folders. To get to the hosts, you could use the path `/**/host/**`.
|
||
|
Note: This may run slowly in very large environments, since a large number of
|
||
|
nodes will be traversed.
|
||
|
|
||
|
## Performance Considerations
|
||
|
|
||
|
### Realtime vs. Historical Metrics
|
||
|
|
||
|
vCenter keeps two different kinds of metrics, known as realtime and historical
|
||
|
metrics.
|
||
|
|
||
|
* Realtime metrics: Available at a 20 second granularity. These metrics are
|
||
|
stored in memory and are very fast and cheap to query. Our tests have shown
|
||
|
that a complete set of realtime metrics for 7000 virtual machines can be
|
||
|
obtained in less than 20 seconds. Realtime metrics are only available on
|
||
|
**ESXi hosts** and **virtual machine** resources. Realtime metrics are only
|
||
|
stored for 1 hour in vCenter.
|
||
|
* Historical metrics: Available at a (default) 5 minute, 30 minutes, 2 hours
|
||
|
and 24 hours rollup levels. The vSphere Telegraf plugin only uses the most
|
||
|
granular rollup which defaults to 5 minutes but can be changed in vCenter to
|
||
|
other interval durations. These metrics are stored in the vCenter database and
|
||
|
can be expensive and slow to query. Historical metrics are the only type of
|
||
|
metrics available for **clusters**, **datastores**, **resource pools** and
|
||
|
**datacenters**.
|
||
|
|
||
|
This distinction has an impact on how Telegraf collects metrics. A single
|
||
|
instance of an input plugin can have one and only one collection interval,
|
||
|
which means that you typically set the collection interval based on the most
|
||
|
frequently collected metric. Let's assume you set the collection interval to 1
|
||
|
minute. All realtime metrics will be collected every minute. Since the
|
||
|
historical metrics are only available on a 5 minute interval, the vSphere
|
||
|
Telegraf plugin automatically skips four out of five collection cycles for
|
||
|
these metrics. This works fine in many cases. Problems arise when the
|
||
|
collection of historical metrics takes longer than the collection interval.
|
||
|
This will cause error messages similar to this to appear in the Telegraf logs:
|
||
|
|
||
|
```text
|
||
|
2019-01-16T13:41:10Z W! [agent] input "inputs.vsphere" did not complete within its interval
|
||
|
```
|
||
|
|
||
|
This will disrupt the metric collection and can result in missed samples. The
|
||
|
best practice workaround is to specify two instances of the vSphere plugin, one
|
||
|
for the realtime metrics with a short collection interval and one for the
|
||
|
historical metrics with a longer interval. You can use the `*_metric_exclude`
|
||
|
to turn off the resources you don't want to collect metrics for in each
|
||
|
instance. For example:
|
||
|
|
||
|
```toml
|
||
|
## Realtime instance
|
||
|
[[inputs.vsphere]]
|
||
|
interval = "60s"
|
||
|
vcenters = [ "https://someaddress/sdk" ]
|
||
|
username = "someuser@vsphere.local"
|
||
|
password = "secret"
|
||
|
|
||
|
insecure_skip_verify = true
|
||
|
force_discover_on_init = true
|
||
|
|
||
|
# Exclude all historical metrics
|
||
|
datastore_metric_exclude = ["*"]
|
||
|
cluster_metric_exclude = ["*"]
|
||
|
datacenter_metric_exclude = ["*"]
|
||
|
resource_pool_metric_exclude = ["*"]
|
||
|
vsan_metric_exclude = ["*"]
|
||
|
|
||
|
collect_concurrency = 5
|
||
|
discover_concurrency = 5
|
||
|
|
||
|
# Historical instance
|
||
|
[[inputs.vsphere]]
|
||
|
|
||
|
interval = "300s"
|
||
|
|
||
|
vcenters = [ "https://someaddress/sdk" ]
|
||
|
username = "someuser@vsphere.local"
|
||
|
password = "secret"
|
||
|
|
||
|
insecure_skip_verify = true
|
||
|
force_discover_on_init = true
|
||
|
host_metric_exclude = ["*"] # Exclude realtime metrics
|
||
|
vm_metric_exclude = ["*"] # Exclude realtime metrics
|
||
|
|
||
|
max_query_metrics = 256
|
||
|
collect_concurrency = 3
|
||
|
```
|
||
|
|
||
|
### Configuring max_query_metrics Setting
|
||
|
|
||
|
The `max_query_metrics` determines the maximum number of metrics to attempt to
|
||
|
retrieve in one call to vCenter. Generally speaking, a higher number means
|
||
|
faster and more efficient queries. However, the number of allowed metrics in a
|
||
|
query is typically limited in vCenter by the `config.vpxd.stats.maxQueryMetrics`
|
||
|
setting in vCenter. The value defaults to 64 on vSphere 5.5 and earlier and to
|
||
|
256 on more recent versions. The vSphere plugin always checks this setting and
|
||
|
will automatically reduce the number if the limit configured in vCenter is lower
|
||
|
than max_query_metrics in the plugin. This will result in a log message similar
|
||
|
to this:
|
||
|
|
||
|
```text
|
||
|
2019-01-21T03:24:18Z W! [input.vsphere] Configured max_query_metrics is 256, but server limits it to 64. Reducing.
|
||
|
```
|
||
|
|
||
|
You may ask a vCenter administrator to increase this limit to help boost
|
||
|
performance.
|
||
|
|
||
|
### Cluster Metrics and the max_query_metrics Setting
|
||
|
|
||
|
Cluster metrics are handled a bit differently by vCenter. They are aggregated
|
||
|
from ESXi and virtual machine metrics and may not be available when you query
|
||
|
their most recent values. When this happens, vCenter will attempt to perform
|
||
|
that aggregation on the fly. Unfortunately, all the subqueries needed
|
||
|
internally in vCenter to perform this aggregation will count towards
|
||
|
`config.vpxd.stats.maxQueryMetrics`. This means that even a very small query
|
||
|
may result in an error message similar to this:
|
||
|
|
||
|
```text
|
||
|
2018-11-02T13:37:11Z E! Error in plugin [inputs.vsphere]: ServerFaultCode: This operation is restricted by the administrator - 'vpxd.stats.maxQueryMetrics'. Contact your system administrator
|
||
|
```
|
||
|
|
||
|
There are two ways of addressing this:
|
||
|
|
||
|
* Ask your vCenter administrator to set `config.vpxd.stats.maxQueryMetrics` to
|
||
|
a number that's higher than the total number of virtual machines managed by a
|
||
|
vCenter instance.
|
||
|
* Exclude the cluster metrics and use either the basicstats aggregator to
|
||
|
calculate sums and averages per cluster or use queries in the visualization
|
||
|
tool to obtain the same result.
|
||
|
|
||
|
### Concurrency Settings
|
||
|
|
||
|
The vSphere plugin allows you to specify two concurrency settings:
|
||
|
|
||
|
* `collect_concurrency`: The maximum number of simultaneous queries for
|
||
|
performance metrics allowed per resource type.
|
||
|
* `discover_concurrency`: The maximum number of simultaneous queries for
|
||
|
resource discovery allowed.
|
||
|
|
||
|
While a higher level of concurrency typically has a positive impact on
|
||
|
performance, increasing these numbers too much can cause performance issues at
|
||
|
the vCenter server. A rule of thumb is to set these parameters to the number of
|
||
|
virtual machines divided by 1500 and rounded up to the nearest integer.
|
||
|
|
||
|
### Configuring historical_interval Setting
|
||
|
|
||
|
When the vSphere plugin queries vCenter for historical statistics it queries for
|
||
|
statistics that exist at a specific interval. The default historical interval
|
||
|
duration is 5 minutes but if this interval has been changed then you must
|
||
|
override the default query interval in the vSphere plugin.
|
||
|
|
||
|
* `historical_interval`: The interval of the most granular statistics configured
|
||
|
in vSphere represented in seconds.
|
||
|
|
||
|
## Metrics
|
||
|
|
||
|
* Cluster Stats
|
||
|
* Cluster services: CPU, memory, failover
|
||
|
* CPU: total, usage
|
||
|
* Memory: consumed, total, vmmemctl
|
||
|
* VM operations: # changes, clone, create, deploy, destroy, power, reboot,
|
||
|
reconfigure, register, reset, shutdown, standby, vmotion
|
||
|
* Host Stats:
|
||
|
* CPU: total, usage, cost, mhz
|
||
|
* Datastore: iops, latency, read/write bytes, # reads/writes
|
||
|
* Disk: commands, latency, kernel reads/writes, # reads/writes, queues
|
||
|
* Memory: total, usage, active, latency, swap, shared, vmmemctl
|
||
|
* Network: broadcast, bytes, dropped, errors, multicast, packets, usage
|
||
|
* Power: energy, usage, capacity
|
||
|
* Res CPU: active, max, running
|
||
|
* Storage Adapter: commands, latency, # reads/writes
|
||
|
* Storage Path: commands, latency, # reads/writes
|
||
|
* System Resources: cpu active, cpu max, cpu running, cpu usage, mem allocated,
|
||
|
mem consumed, mem shared, swap
|
||
|
* System: uptime
|
||
|
* Flash Module: active VMDKs
|
||
|
* VM Stats:
|
||
|
* CPU: demand, usage, readiness, cost, mhz
|
||
|
* Datastore: latency, # reads/writes
|
||
|
* Disk: commands, latency, # reads/writes, provisioned, usage
|
||
|
* Memory: granted, usage, active, swap, vmmemctl, memorySizeMB (allocated),
|
||
|
memoryReservation
|
||
|
* Network: broadcast, bytes, dropped, multicast, packets, usage
|
||
|
* Power: energy, usage
|
||
|
* Res CPU: active, max, running
|
||
|
* System: operating system uptime, uptime
|
||
|
* Virtual Disk: seeks, # reads/writes, latency, load
|
||
|
* Resource Pools stats:
|
||
|
* Memory: total, usage, active, latency, swap, shared, vmmemctl
|
||
|
* CPU: capacity, usage, corecount
|
||
|
* Disk: throughput
|
||
|
* Network: throughput
|
||
|
* Power: energy, usage
|
||
|
* Datastore stats:
|
||
|
* Disk: Capacity, provisioned, used
|
||
|
* Numeric Sensor stats:
|
||
|
* CPU: temperature
|
||
|
|
||
|
For a detailed list of commonly available metrics, please refer to
|
||
|
[METRICS.md](METRICS.md)
|
||
|
|
||
|
### Tags
|
||
|
|
||
|
* all metrics
|
||
|
* vcenter (vcenter url)
|
||
|
* all host metrics
|
||
|
* cluster (vcenter cluster)
|
||
|
* all vm metrics
|
||
|
* cluster (vcenter cluster)
|
||
|
* esxhost (name of ESXi host)
|
||
|
* guest (guest operating system id)
|
||
|
* resource pool (name of resource pool)
|
||
|
* cpu stats for Host and VM
|
||
|
* cpu (cpu core - not all CPU fields will have this tag)
|
||
|
* datastore stats for Host and VM
|
||
|
* datastore (id of datastore)
|
||
|
* disk stats for Host and VM
|
||
|
* disk (name of disk)
|
||
|
* disk.used.capacity for Datastore
|
||
|
* disk (type of disk)
|
||
|
* net stats for Host and VM
|
||
|
* interface (name of network interface)
|
||
|
* storageAdapter stats for Host
|
||
|
* adapter (name of storage adapter)
|
||
|
* storagePath stats for Host
|
||
|
* path (id of storage path)
|
||
|
* sys.resource* stats for Host
|
||
|
* resource (resource type)
|
||
|
* vflashModule stats for Host
|
||
|
* module (name of flash module)
|
||
|
* virtualDisk stats for VM
|
||
|
* disk (name of virtual disk)
|
||
|
|
||
|
## Add a vSAN extension
|
||
|
|
||
|
A vSAN resource is a special type of resource that can be collected by the
|
||
|
plugin. The configuration of a vSAN resource slightly differs from the
|
||
|
configuration of hosts, VMs, and other resources.
|
||
|
|
||
|
### Prerequisites for vSAN
|
||
|
|
||
|
* vSphere 6.5 and later
|
||
|
* Clusters with vSAN enabled
|
||
|
* [Turn on Virtual SAN performance service][vsan_perf_service]: When you create
|
||
|
a vSAN cluster, the performance service is disabled. To monitor the
|
||
|
performance metrics, you must turn on vSAN performance service.
|
||
|
|
||
|
[vsan_perf_service]: https://docs.vmware.com/en/VMware-vSphere/6.5/com.vmware.vsphere.virtualsan.doc/GUID-02F67DC3-3D5A-48A4-A445-D2BD6AF2862C.html
|
||
|
|
||
|
### vSAN Configuration
|
||
|
|
||
|
```toml
|
||
|
[[inputs.vsphere]]
|
||
|
interval = "300s"
|
||
|
vcenters = ["https://<vcenter-ip>/sdk", "https://<vcenter2-ip>/sdk"]
|
||
|
username = "<user>"
|
||
|
password = "<pwd>"
|
||
|
|
||
|
# Exclude all other metrics
|
||
|
vm_metric_exclude = ["*"]
|
||
|
datastore_metric_exclude = ["*"]
|
||
|
datacenter_metric_exclude = ["*"]
|
||
|
host_metric_exclude = ["*"]
|
||
|
cluster_metric_exclude = ["*"]
|
||
|
|
||
|
# By default all supported entity will be included
|
||
|
vsan_metric_include = [
|
||
|
"summary.disk-usage",
|
||
|
"summary.health",
|
||
|
"summary.resync",
|
||
|
"performance.cluster-domclient",
|
||
|
"performance.cluster-domcompmgr",
|
||
|
"performance.host-domclient",
|
||
|
"performance.host-domcompmgr",
|
||
|
"performance.cache-disk",
|
||
|
"performance.disk-group",
|
||
|
"performance.capacity-disk",
|
||
|
"performance.disk-group",
|
||
|
"performance.virtual-machine",
|
||
|
"performance.vscsi",
|
||
|
"performance.virtual-disk",
|
||
|
"performance.vsan-host-net",
|
||
|
"performance.vsan-vnic-net",
|
||
|
"performance.vsan-pnic-net",
|
||
|
"performance.vsan-iscsi-host",
|
||
|
"performance.vsan-iscsi-target",
|
||
|
"performance.vsan-iscsi-lun",
|
||
|
"performance.lsom-world-cpu",
|
||
|
"performance.nic-world-cpu",
|
||
|
"performance.dom-world-cpu",
|
||
|
"performance.cmmds-world-cpu",
|
||
|
"performance.host-cpu",
|
||
|
"performance.host-domowner",
|
||
|
"performance.host-memory-slab",
|
||
|
"performance.host-memory-heap",
|
||
|
"performance.system-mem",
|
||
|
]
|
||
|
# by default vsan_metric_skip_verify = false
|
||
|
vsan_metric_skip_verify = true
|
||
|
vsan_metric_exclude = [ ]
|
||
|
# vsan_cluster_include = [ "/*/host/**" ] # Inventory path to clusters to collect (by default all are collected)
|
||
|
|
||
|
collect_concurrency = 5
|
||
|
discover_concurrency = 5
|
||
|
|
||
|
## Optional SSL Config
|
||
|
# ssl_ca = "/path/to/cafile"
|
||
|
# ssl_cert = "/path/to/certfile"
|
||
|
# ssl_key = "/path/to/keyfile"
|
||
|
## Use SSL but skip chain & host verification
|
||
|
# insecure_skip_verify = false
|
||
|
```
|
||
|
|
||
|
Use `vsan_metric_include = [...]` to define the vSAN metrics that you want to
|
||
|
collect. For example
|
||
|
|
||
|
```toml
|
||
|
vsan_metric_include = ["summary.*", "performance.host-domclient", "performance.cache-disk", "performance.disk-group", "performance.capacity-disk"]
|
||
|
```
|
||
|
|
||
|
To include all supported vSAN metrics, use `vsan_metric_include = [ "*" ]`.
|
||
|
To disable all the vSAN metrics, use `vsan_metric_exclude = [ "*" ]`.
|
||
|
|
||
|
`vsan_metric_skip_verify` defines whether to skip verifying vSAN metrics against
|
||
|
the ones from [GetSupportedEntityTypes API][supported_entity_types]. This option
|
||
|
is given because some performance entities are not returned by the API, but we
|
||
|
want to offer the flexibility if you really need the stats. When set to false,
|
||
|
anything not in the supported entity list will be filtered out. When set to
|
||
|
true, queried metrics will be identical to vsan_metric_include and the exclusive
|
||
|
array will not be used in this case. By default the value is false.
|
||
|
|
||
|
`vsan_cluster_include` defines a list of inventory paths that will be used to
|
||
|
select a portion of vSAN clusters. vSAN metrics are only collected on the
|
||
|
cluster level. Therefore, use the same way as inventory paths for
|
||
|
[vSphere clusters](#inventory-paths).
|
||
|
|
||
|
Many vCenter environments use self-signed certificates. Update the bottom
|
||
|
portion of the above configuration and provide proper values for all applicable
|
||
|
SSL Config settings that apply in your vSphere environment. In some
|
||
|
environments, setting insecure_skip_verify = true will be necessary when the
|
||
|
SSL certificates are not available.
|
||
|
|
||
|
To ensure consistent collection in larger vSphere environments, you must
|
||
|
increase concurrency for the plugin. Use the collect_concurrency setting to
|
||
|
control concurrency. Set collect_concurrency to the number of virtual machines
|
||
|
divided by 1500 and rounded up to the nearest integer. For example, for
|
||
|
1200 VMs use 1, and for 2300 VMs use 2.
|
||
|
|
||
|
[supported_entity_types]: https://code.vmware.com/apis/48/vsan#/doc/vim.cluster.VsanPerformanceManager.html#getSupportedEntityTypes
|
||
|
|
||
|
### Measurements & Fields
|
||
|
|
||
|
**NOTE**: Depending on the vSAN version, the vSAN performance measurements
|
||
|
and fields may vary.
|
||
|
|
||
|
* vSAN Summary
|
||
|
* overall_health
|
||
|
* total_capacity_bytes, free_capacity_bytes
|
||
|
* total_bytes_to_sync, total_objects_to_sync, total_recovery_eta
|
||
|
|
||
|
* vSAN Performance
|
||
|
* cluster-domclient
|
||
|
* iops_read, throughput_read, latency_avg_read, iops_write,
|
||
|
throughput_write, latency_avg_write, congestion, oio
|
||
|
* cluster-domcompmgr
|
||
|
* iops_read, throughput_read, latency_avg_read, iops_write,
|
||
|
throughput_write, latency_avg_write, iops_rec_write, throughput_rec_write,
|
||
|
latency_avg_rec_write, congestion, oio, iops_resync_read, tput_resync_read,
|
||
|
lat_avg_resyncread
|
||
|
* host-domclient
|
||
|
* iops_read, throughput_read, latency_avg_read, read_count, iops_write,
|
||
|
throughput_write, latency_avg_write, write_count, congestion, oio,
|
||
|
client_cache_hits, client_cache_hit_rate
|
||
|
* host-domcompmgr
|
||
|
* iops_read, throughput_read, latency_avg_read, read_count, iops_write,
|
||
|
throughput_write, latency_avg_write, write_count, iops_rec_write,
|
||
|
throughput_rec_write, latency_avg_rec_write, rec_write_count congestion,
|
||
|
oio, iops_resync_read, tput_resync_read, lat_avg_resync_read
|
||
|
* cache-disk
|
||
|
* iops_dev_read, throughput_dev_read, latency_dev_read, io_count_dev_read,
|
||
|
iops_dev_write, throughput_dev_write, latency_dev_write,
|
||
|
io_count_dev_write, latency_dev_d_avg, latency_dev_g_avg
|
||
|
* capacity-disk
|
||
|
* iops_dev_read, throughput_dev_read, latency_dev_read, io_count_dev_read,
|
||
|
iops_dev_write, throughput_dev_write, latency_dev_write,
|
||
|
io_count_dev_write, latency_dev_d_avg, latency_dev_g_avg, iops_read,
|
||
|
latency_read, io_count_read, iops_write, latency_write, io_count_write
|
||
|
* disk-group
|
||
|
* iops_sched, latency_sched, outstanding_bytes_sched, iops_sched_queue_rec,
|
||
|
throughput_sched_queue_rec,latency_sched_queue_rec, iops_sched_queue_vm,
|
||
|
throughput_sched_queue_vm,latency_sched_queue_vm, iops_sched_queue_meta,
|
||
|
throughput_sched_queue_meta,latency_sched_queue_meta,
|
||
|
iops_delay_pct_sched, latency_delay_sched, rc_hit_rate, wb_free_pct,
|
||
|
war_evictions, quota_evictions, iops_rc_read, latency_rc_read,
|
||
|
io_count_rc_read, iops_wb_read, latency_wb_read, io_count_wb_read,
|
||
|
iops_rc_write, latency_rc_write, io_count_rc_write, iops_wb_write,
|
||
|
latency_wb_write, io_count_wb_write, ssd_bytes_drained, zero_bytes_drained,
|
||
|
mem_congestion, slab_congestion, ssd_congestion, iops_congestion,
|
||
|
log_congestion, comp_congestion, iops_direct_sched, iops_read,
|
||
|
throughput_read, latency_avg_read, read_count, iops_write,
|
||
|
throughput_write, latency_avg_write, write_count, oio_write,
|
||
|
oio_rec_write, oio_write_size, oio_rec_write_size, rc_size, wb_size,
|
||
|
capacity, capacity_used, capacity_reserved, throughput_sched,
|
||
|
iops_resync_read_policy, iops_resync_read_decom,
|
||
|
iops_resync_read_rebalance, iops_resync_read_fix_comp,
|
||
|
iops_resync_write_policy, iops_resync_write_decom,
|
||
|
iops_resync_write_rebalance, iops_resync_write_fix_comp,
|
||
|
tput_resync_read_policy, tput_resync_read_decom,
|
||
|
tput_resync_read_rebalance, tput_resync_read_fix_comp,
|
||
|
tput_resync_write_policy, tput_resync_write_decom,
|
||
|
tput_resync_write_rebalance, tput_resync_write_fix_comp,
|
||
|
lat_resync_read_policy, lat_resync_read_decom, lat_resync_read_rebalance,
|
||
|
lat_resync_read_fix_comp, lat_resync_write_policy, lat_resync_write_decom,
|
||
|
lat_resync_write_rebalance, lat_resync_write_fix_comp
|
||
|
* virtual-machine
|
||
|
* iops_read, throughput_read, latency_read_avg, latency_read_stddev,
|
||
|
read_count, iops_write, throughput_write, latency_write_avg,
|
||
|
latency_write_stddev, write_count
|
||
|
* vscsi
|
||
|
* iops_read, throughput_read, latency_read, read_count, iops_write,
|
||
|
throughput_write, latency_write, write_count
|
||
|
* virtual-disk
|
||
|
* iops_limit, niops, niops_delayed
|
||
|
* vsan-host-net
|
||
|
* rx_throughput, rx_packets, rx_packets_loss_rate, tx_throughput, tx_packets,
|
||
|
tx_packets_loss_rate
|
||
|
* vsan-vnic-net
|
||
|
* rx_throughput, rx_packets, rx_packets_loss_rate, tx_throughput, tx_packets,
|
||
|
tx_packets_loss_rate
|
||
|
* vsan-pnic-net
|
||
|
* rx_throughput, rx_packets, rx_packets_loss_rate, tx_throughput, tx_packets,
|
||
|
tx_packets_loss_rate
|
||
|
* vsan-iscsi-host
|
||
|
* iops_read, iops_write, iops_total, bandwidth_read, bandwidth_write,
|
||
|
bandwidth_total, latency_read, latency_write, latency_total, queue_depth
|
||
|
* vsan-iscsi-target
|
||
|
* iops_read, iops_write, iops_total, bandwidth_read, bandwidth_write,
|
||
|
bandwidth_total, latency_read, latency_write, latency_total, queue_depth
|
||
|
* vsan-iscsi-lun
|
||
|
* iops_read, iops_write, iops_total, bandwidth_read, bandwidth_write,
|
||
|
bandwidth_total, latency_read, latency_write, latency_total, queue_depth
|
||
|
|
||
|
### vSAN Tags
|
||
|
|
||
|
* all vSAN metrics
|
||
|
* vcenter
|
||
|
* dcname
|
||
|
* clustername
|
||
|
* moid (the cluster's managed object id)
|
||
|
* host-domclient, host-domcompmgr
|
||
|
* hostname
|
||
|
* disk-group, cache-disk, capacity-disk
|
||
|
* hostname
|
||
|
* deviceName
|
||
|
* ssdUuid (if SSD)
|
||
|
* vsan-host-net
|
||
|
* hostname
|
||
|
* vsan-pnic-net
|
||
|
* pnic
|
||
|
* vsan-vnic-net
|
||
|
* vnic
|
||
|
* stackName
|
||
|
|
||
|
### Realtime vs. Historical Metrics in vSAN
|
||
|
|
||
|
vSAN metrics also keep two different kinds of metrics - realtime and
|
||
|
historical metrics.
|
||
|
|
||
|
* Realtime metrics are metrics with the prefix 'summary'. These metrics are
|
||
|
available in realtime.
|
||
|
* Historical metrics are metrics with the prefix 'performance'. These are
|
||
|
metrics queried from vSAN performance API, which is available at a 5-minute
|
||
|
rollup level.
|
||
|
|
||
|
For performance consideration, it is better to specify two instances of the
|
||
|
plugin, one for the realtime metrics with a short collection interval,
|
||
|
and the second one - for the historical metrics with a longer interval.
|
||
|
For example:
|
||
|
|
||
|
```toml
|
||
|
## Realtime instance
|
||
|
[[inputs.vsphere]]
|
||
|
interval = "30s"
|
||
|
vcenters = [ "https://someaddress/sdk" ]
|
||
|
username = "someuser@vsphere.local"
|
||
|
password = "secret"
|
||
|
|
||
|
insecure_skip_verify = true
|
||
|
force_discover_on_init = true
|
||
|
|
||
|
# Exclude all other metrics
|
||
|
vm_metric_exclude = ["*"]
|
||
|
datastore_metric_exclude = ["*"]
|
||
|
datacenter_metric_exclude = ["*"]
|
||
|
host_metric_exclude = ["*"]
|
||
|
cluster_metric_exclude = ["*"]
|
||
|
|
||
|
vsan_metric_include = [ "summary.*" ]
|
||
|
vsan_metric_exclude = [ ]
|
||
|
vsan_metric_skip_verify = false
|
||
|
|
||
|
collect_concurrency = 5
|
||
|
discover_concurrency = 5
|
||
|
|
||
|
# Historical instance
|
||
|
[[inputs.vsphere]]
|
||
|
|
||
|
interval = "300s"
|
||
|
vcenters = [ "https://someaddress/sdk" ]
|
||
|
username = "someuser@vsphere.local"
|
||
|
password = "secret"
|
||
|
|
||
|
insecure_skip_verify = true
|
||
|
force_discover_on_init = true
|
||
|
|
||
|
# Exclude all other metrics
|
||
|
vm_metric_exclude = ["*"]
|
||
|
datastore_metric_exclude = ["*"]
|
||
|
datacenter_metric_exclude = ["*"]
|
||
|
host_metric_exclude = ["*"]
|
||
|
cluster_metric_exclude = ["*"]
|
||
|
|
||
|
vsan_metric_include = [ "performance.*" ]
|
||
|
vsan_metric_exclude = [ ]
|
||
|
vsan_metric_skip_verify = false
|
||
|
|
||
|
collect_concurrency = 5
|
||
|
discover_concurrency = 5
|
||
|
```
|
||
|
|
||
|
## Example Output
|
||
|
|
||
|
```text
|
||
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 run_summation=2608i,ready_summation=129i,usage_average=5.01,used_summation=2134i,demand_average=326i 1535660299000000000
|
||
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 bytesRx_average=321i,bytesTx_average=335i 1535660299000000000
|
||
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 write_average=144i,read_average=4i 1535660299000000000
|
||
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 bytesRx_average=242i,bytesTx_average=308i 1535660299000000000
|
||
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 write_average=232i,read_average=4i 1535660299000000000
|
||
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 usage_average=5.49,used_summation=1804i,demand_average=308i,run_summation=2001i,ready_summation=120i 1535660299000000000
|
||
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 usage_average=4.19,used_summation=2108i,demand_average=285i,run_summation=1793i,ready_summation=93i 1535660299000000000
|
||
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 bytesRx_average=272i,bytesTx_average=419i 1535660299000000000
|
||
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 write_average=229i,read_average=4i 1535660299000000000
|
||
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 run_summation=2277i,ready_summation=118i,usage_average=4.67,used_summation=2546i,demand_average=289i 1535660299000000000
|
||
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 bytesRx_average=243i,bytesTx_average=296i 1535660299000000000
|
||
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 write_average=158i,read_average=4i 1535660299000000000
|
||
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,interface=vmnic0,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=1042i,bytesTx_average=753i,bytesRx_average=660i 1535660299000000000
|
||
|
vsphere_host_cpu,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 utilization_average=10.46,usage_average=22.4,readiness_average=0.4,costop_summation=2i,coreUtilization_average=19.61,wait_summation=5148518i,idle_summation=58581i,latency_average=0.6,ready_summation=13370i,used_summation=19219i 1535660299000000000
|
||
|
vsphere_host_cpu,cpu=0,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=25.6,utilization_average=11.58,used_summation=24306i,usage_average=24.26,idle_summation=86688i 1535660299000000000
|
||
|
vsphere_host_cpu,cpu=1,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=12.29,utilization_average=8.32,used_summation=31312i,usage_average=22.47,idle_summation=94934i 1535660299000000000
|
||
|
vsphere_host_disk,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 read_average=331i,write_average=2800i 1535660299000000000
|
||
|
vsphere_host_disk,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 write_average=2701i,read_average=258i 1535660299000000000
|
||
|
vsphere_host_mem,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=93.27 1535660299000000000
|
||
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 bytesTx_average=650i,usage_average=1414i,bytesRx_average=569i 1535660299000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,cpu=1,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 utilization_average=12.6,used_summation=25775i,usage_average=24.44,idle_summation=68886i,coreUtilization_average=17.59 1535660299000000000
|
||
|
vsphere_host_disk,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 read_average=340i,write_average=2340i 1535660299000000000
|
||
|
vsphere_host_disk,clustername=DC0_C0,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 write_average=2277i,read_average=282i 1535660299000000000
|
||
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=104.78 1535660299000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesTx_average=463i,usage_average=1131i,bytesRx_average=719i 1535660299000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,interface=vmnic0,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=1668i,bytesTx_average=838i,bytesRx_average=921i 1535660299000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 used_summation=28952i,utilization_average=11.36,idle_summation=93261i,latency_average=0.46,ready_summation=12837i,usage_average=21.56,readiness_average=0.39,costop_summation=2i,coreUtilization_average=27.19,wait_summation=3820829i 1535660299000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,cpu=0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 coreUtilization_average=24.12,utilization_average=13.83,used_summation=22462i,usage_average=24.69,idle_summation=96993i 1535660299000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=localhost:8989 connect_ns=4727607i,discover_ns=65389011i,discovered_objects=8i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=localhost:8989 gather_duration_ns=296223i,gather_count=0i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=192.168.1.151 gather_duration_ns=136050i,gather_count=0i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=localhost:8989 gather_count=62i,gather_duration_ns=8788033i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=192.168.1.151 gather_count=0i,gather_duration_ns=162002i 1535660309000000000
|
||
|
internal_gather,host=host.example.com,input=vsphere,os=Mac gather_time_ns=17483653i,metrics_gathered=28i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=192.168.1.151 connect_ns=0i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=localhost:8989 gather_duration_ns=7291897i,gather_count=36i 1535660309000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=192.168.1.151 gather_duration_ns=958474i,gather_count=0i 1535660309000000000
|
||
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 usage_average=8.82,used_summation=3192i,demand_average=283i,run_summation=2419i,ready_summation=115i 1535660319000000000
|
||
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 bytesRx_average=277i,bytesTx_average=343i 1535660319000000000
|
||
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 read_average=1i,write_average=741i 1535660319000000000
|
||
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 bytesRx_average=386i,bytesTx_average=369i 1535660319000000000
|
||
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 write_average=814i,read_average=1i 1535660319000000000
|
||
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 run_summation=1778i,ready_summation=111i,usage_average=7.54,used_summation=2339i,demand_average=297i 1535660319000000000
|
||
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 usage_average=6.98,used_summation=2125i,demand_average=211i,run_summation=2990i,ready_summation=141i 1535660319000000000
|
||
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 bytesRx_average=357i,bytesTx_average=268i 1535660319000000000
|
||
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 write_average=528i,read_average=1i 1535660319000000000
|
||
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 used_summation=2374i,demand_average=195i,run_summation=3454i,ready_summation=110i,usage_average=7.34 1535660319000000000
|
||
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 bytesRx_average=308i,bytesTx_average=246i 1535660319000000000
|
||
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 write_average=1178i,read_average=1i 1535660319000000000
|
||
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,interface=vmnic0,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 bytesRx_average=773i,usage_average=1521i,bytesTx_average=890i 1535660319000000000
|
||
|
vsphere_host_cpu,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 wait_summation=3421258i,idle_summation=67994i,latency_average=0.36,usage_average=29.86,readiness_average=0.37,used_summation=25244i,costop_summation=2i,coreUtilization_average=21.94,utilization_average=17.19,ready_summation=15897i 1535660319000000000
|
||
|
vsphere_host_cpu,cpu=0,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 utilization_average=11.32,used_summation=19333i,usage_average=14.29,idle_summation=92708i,coreUtilization_average=27.68 1535660319000000000
|
||
|
vsphere_host_cpu,cpu=1,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 used_summation=28596i,usage_average=25.32,idle_summation=79553i,coreUtilization_average=28.01,utilization_average=11.33 1535660319000000000
|
||
|
vsphere_host_disk,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 read_average=86i,write_average=1659i 1535660319000000000
|
||
|
vsphere_host_disk,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 write_average=1997i,read_average=58i 1535660319000000000
|
||
|
vsphere_host_mem,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=68.45 1535660319000000000
|
||
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 bytesTx_average=679i,usage_average=2286i,bytesRx_average=719i 1535660319000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,cpu=1,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 utilization_average=10.52,used_summation=21693i,usage_average=23.09,idle_summation=84590i,coreUtilization_average=29.92 1535660319000000000
|
||
|
vsphere_host_disk,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 read_average=113i,write_average=1236i 1535660319000000000
|
||
|
vsphere_host_disk,clustername=DC0_C0,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 write_average=1708i,read_average=110i 1535660319000000000
|
||
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=111.46 1535660319000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesTx_average=998i,usage_average=2000i,bytesRx_average=881i 1535660319000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,interface=vmnic0,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=1683i,bytesTx_average=675i,bytesRx_average=1078i 1535660319000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 used_summation=28531i,wait_summation=3139129i,utilization_average=9.99,idle_summation=98579i,latency_average=0.51,costop_summation=2i,coreUtilization_average=14.35,ready_summation=16121i,usage_average=34.19,readiness_average=0.4 1535660319000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,cpu=0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 utilization_average=12.2,used_summation=22750i,usage_average=18.84,idle_summation=99539i,coreUtilization_average=23.05 1535660319000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=localhost:8989 gather_duration_ns=7076543i,gather_count=62i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=host,vcenter=192.168.1.151 gather_duration_ns=4051303i,gather_count=0i 1535660339000000000
|
||
|
internal_gather,host=host.example.com,input=vsphere,os=Mac metrics_gathered=56i,gather_time_ns=13555029i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=192.168.1.151 connect_ns=0i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=localhost:8989 gather_duration_ns=6335467i,gather_count=36i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=192.168.1.151 gather_duration_ns=958474i,gather_count=0i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,vcenter=localhost:8989 discover_ns=65389011i,discovered_objects=8i,connect_ns=4727607i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=datastore,vcenter=localhost:8989 gather_duration_ns=296223i,gather_count=0i 1535660339000000000
|
||
|
internal_vsphere,host=host.example.com,os=Mac,resourcetype=vm,vcenter=192.168.1.151 gather_count=0i,gather_duration_ns=1540920i 1535660339000000000
|
||
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 write_average=302i,read_average=11i 1535660339000000000
|
||
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 usage_average=5.58,used_summation=2941i,demand_average=298i,run_summation=3255i,ready_summation=96i 1535660339000000000
|
||
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-35,os=Mac,source=DC0_H0_VM0,vcenter=localhost:8989,vmname=DC0_H0_VM0 bytesRx_average=155i,bytesTx_average=241i 1535660339000000000
|
||
|
vsphere_vm_cpu,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 usage_average=10.3,used_summation=3053i,demand_average=346i,run_summation=3289i,ready_summation=122i 1535660339000000000
|
||
|
vsphere_vm_net,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 bytesRx_average=215i,bytesTx_average=275i 1535660339000000000
|
||
|
vsphere_vm_virtualDisk,esxhostname=DC0_H0,guest=other,host=host.example.com,moid=vm-38,os=Mac,source=DC0_H0_VM1,vcenter=localhost:8989,vmname=DC0_H0_VM1 write_average=252i,read_average=14i 1535660339000000000
|
||
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 usage_average=8,used_summation=2183i,demand_average=354i,run_summation=3542i,ready_summation=128i 1535660339000000000
|
||
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 bytesRx_average=178i,bytesTx_average=200i 1535660339000000000
|
||
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-41,os=Mac,source=DC0_C0_RP0_VM0,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM0 write_average=283i,read_average=12i 1535660339000000000
|
||
|
vsphere_vm_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 demand_average=328i,run_summation=3481i,ready_summation=122i,usage_average=7.95,used_summation=2167i 1535660339000000000
|
||
|
vsphere_vm_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 bytesTx_average=282i,bytesRx_average=196i 1535660339000000000
|
||
|
vsphere_vm_virtualDisk,clustername=DC0_C0,esxhostname=DC0_C0_H0,guest=other,host=host.example.com,moid=vm-44,os=Mac,source=DC0_C0_RP0_VM1,vcenter=localhost:8989,vmname=DC0_C0_RP0_VM1 write_average=321i,read_average=13i 1535660339000000000
|
||
|
vsphere_host_disk,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 read_average=39i,write_average=2635i 1535660339000000000
|
||
|
vsphere_host_disk,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 write_average=2635i,read_average=30i 1535660339000000000
|
||
|
vsphere_host_mem,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=98.5 1535660339000000000
|
||
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=1887i,bytesRx_average=662i,bytesTx_average=251i 1535660339000000000
|
||
|
vsphere_host_net,esxhostname=DC0_H0,host=host.example.com,interface=vmnic0,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 usage_average=1481i,bytesTx_average=899i,bytesRx_average=992i 1535660339000000000
|
||
|
vsphere_host_cpu,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 used_summation=50405i,costop_summation=2i,utilization_average=17.32,latency_average=0.61,ready_summation=14843i,usage_average=27.94,coreUtilization_average=32.12,wait_summation=3058787i,idle_summation=56600i,readiness_average=0.36 1535660339000000000
|
||
|
vsphere_host_cpu,cpu=0,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=37.61,utilization_average=17.05,used_summation=38013i,usage_average=32.66,idle_summation=89575i 1535660339000000000
|
||
|
vsphere_host_cpu,cpu=1,esxhostname=DC0_H0,host=host.example.com,moid=host-19,os=Mac,source=DC0_H0,vcenter=localhost:8989 coreUtilization_average=25.92,utilization_average=18.72,used_summation=39790i,usage_average=40.42,idle_summation=69457i 1535660339000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,interface=vmnic0,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=1246i,bytesTx_average=673i,bytesRx_average=781i 1535660339000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 coreUtilization_average=33.8,idle_summation=77121i,ready_summation=15857i,readiness_average=0.39,used_summation=29554i,costop_summation=2i,wait_summation=4338417i,utilization_average=17.87,latency_average=0.44,usage_average=28.78 1535660339000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,cpu=0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 idle_summation=86610i,coreUtilization_average=34.36,utilization_average=19.03,used_summation=28766i,usage_average=23.72 1535660339000000000
|
||
|
vsphere_host_cpu,clustername=DC0_C0,cpu=1,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 coreUtilization_average=33.15,utilization_average=16.8,used_summation=44282i,usage_average=30.08,idle_summation=93490i 1535660339000000000
|
||
|
vsphere_host_disk,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 read_average=56i,write_average=1672i 1535660339000000000
|
||
|
vsphere_host_disk,clustername=DC0_C0,disk=/var/folders/rf/txwdm4pj409f70wnkdlp7sz80000gq/T/govcsim-DC0-LocalDS_0-367088371@folder-5,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 write_average=2110i,read_average=48i 1535660339000000000
|
||
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=116.21 1535660339000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesRx_average=726i,bytesTx_average=643i,usage_average=1504i 1535660339000000000
|
||
|
vsphere_host_mem,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 usage_average=116.21 1535660339000000000
|
||
|
vsphere_host_net,clustername=DC0_C0,esxhostname=DC0_C0_H0,host=host.example.com,moid=host-30,os=Mac,source=DC0_C0_H0,vcenter=localhost:8989 bytesRx_average=726i,bytesTx_average=643i,usage_average=1504i 1535660339000000000
|
||
|
```
|
||
|
|
||
|
## vSAN Sample Output
|
||
|
|
||
|
```text
|
||
|
vsphere_vsan_performance_hostdomclient,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,hostname=DC0_C0_H0,moid=domain-c8,source=Example-VSAN,vcenter=localhost:8898 iops_read=7,write_congestion=0,unmap_congestion=0,read_count=2199,iops=8,latency_max_write=8964,latency_avg_unmap=0,latency_avg_write=1883,write_count=364,num_oio=12623,throughput=564127,client_cache_hits=0,latency_max_read=17821,latency_max_unmap=0,read_congestion=0,latency_avg=1154,congestion=0,throughput_read=554721,latency_avg_read=1033,throughput_write=9406,client_cache_hit_rate=0,iops_unmap=0,throughput_unmap=0,latency_stddev=1315,io_count=2563,oio=4,iops_write=1,unmap_count=0 1578955200000000000
|
||
|
vsphere_vsan_performance_clusterdomcompmgr,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,uuid=XXXXXXXX-XXXX-XXXX-XXXX-XXXXXXXXX,vcenter=localhost:8898 latency_avg_rec_write=0,latency_avg_write=9886,congestion=0,iops_resync_read=0,lat_avg_resync_read=0,iops_read=289,latency_avg_read=1184,throughput_write=50137368,iops_rec_write=0,throughput_rec_write=0,tput_resync_read=0,throughput_read=9043654,iops_write=1272,oio=97 1578954900000000000
|
||
|
vsphere_vsan_performance_clusterdomclient,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,uuid=XXXXXXXX-XXXX-XXXX-XXXX-XXXXXXXXX,vcenter=localhost:8898 latency_avg_write=1011,congestion=0,oio=26,iops_read=6,throughput_read=489093,latency_avg_read=1085,iops_write=43,throughput_write=435142 1578955200000000000
|
||
|
vsphere_vsan_summary,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,vcenter=localhost:8898 total_bytes_to_sync=0i,total_objects_to_sync=0i,total_recovery_eta=0i 1578955489000000000
|
||
|
vsphere_vsan_summary,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,vcenter=localhost:8898 overall_health=1i 1578955489000000000
|
||
|
vsphere_vsan_summary,clustername=Example-VSAN,dcname=Example-DC,host=host.example.com,moid=domain-c7,source=Example-VSAN,vcenter=localhost:8898 free_capacity_byte=11022535578757i,total_capacity_byte=14102625779712i 1578955488000000000
|
||
|
```
|