Update telegraf.conf
This commit is contained in:
parent
634762b739
commit
96f3d7def4
|
@ -183,8 +183,8 @@
|
||||||
# ## If true, exchange will be passively declared.
|
# ## If true, exchange will be passively declared.
|
||||||
# # exchange_declare_passive = false
|
# # exchange_declare_passive = false
|
||||||
#
|
#
|
||||||
# ## If true, exchange will be created as a durable exchange.
|
# ## Exchange durability can be either "transient" or "durable".
|
||||||
# # exchange_durable = true
|
# # exchange_durability = "durable"
|
||||||
#
|
#
|
||||||
# ## Additional exchange arguments.
|
# ## Additional exchange arguments.
|
||||||
# # exchange_arguments = { }
|
# # exchange_arguments = { }
|
||||||
|
@ -256,7 +256,7 @@
|
||||||
# # timeout = "5s"
|
# # timeout = "5s"
|
||||||
#
|
#
|
||||||
# ## Enable additional diagnostic logging.
|
# ## Enable additional diagnostic logging.
|
||||||
# # enable_diagnosic_logging = false
|
# # enable_diagnostic_logging = false
|
||||||
#
|
#
|
||||||
# ## Context Tag Sources add Application Insights context tags to a tag value.
|
# ## Context Tag Sources add Application Insights context tags to a tag value.
|
||||||
# ##
|
# ##
|
||||||
|
@ -267,6 +267,32 @@
|
||||||
# # "ai.cloud.roleInstance" = "kubernetes_pod_name"
|
# # "ai.cloud.roleInstance" = "kubernetes_pod_name"
|
||||||
|
|
||||||
|
|
||||||
|
# # Send aggregate metrics to Azure Monitor
|
||||||
|
# [[outputs.azure_monitor]]
|
||||||
|
# ## Timeout for HTTP writes.
|
||||||
|
# # timeout = "20s"
|
||||||
|
#
|
||||||
|
# ## Set the namespace prefix, defaults to "Telegraf/<input-name>".
|
||||||
|
# # namespace_prefix = "Telegraf/"
|
||||||
|
#
|
||||||
|
# ## Azure Monitor doesn't have a string value type, so convert string
|
||||||
|
# ## fields to dimensions (a.k.a. tags) if enabled. Azure Monitor allows
|
||||||
|
# ## a maximum of 10 dimensions so Telegraf will only send the first 10
|
||||||
|
# ## alphanumeric dimensions.
|
||||||
|
# # strings_as_dimensions = false
|
||||||
|
#
|
||||||
|
# ## Both region and resource_id must be set or be available via the
|
||||||
|
# ## Instance Metadata service on Azure Virtual Machines.
|
||||||
|
# #
|
||||||
|
# ## Azure Region to publish metrics against.
|
||||||
|
# ## ex: region = "southcentralus"
|
||||||
|
# # region = ""
|
||||||
|
# #
|
||||||
|
# ## The Azure Resource ID against which metric will be logged, e.g.
|
||||||
|
# ## ex: resource_id = "/subscriptions/<subscription_id>/resourceGroups/<resource_group>/providers/Microsoft.Compute/virtualMachines/<vm_name>"
|
||||||
|
# # resource_id = ""
|
||||||
|
|
||||||
|
|
||||||
# # Configuration for AWS CloudWatch output.
|
# # Configuration for AWS CloudWatch output.
|
||||||
# [[outputs.cloudwatch]]
|
# [[outputs.cloudwatch]]
|
||||||
# ## Amazon REGION
|
# ## Amazon REGION
|
||||||
|
@ -287,8 +313,22 @@
|
||||||
# #profile = ""
|
# #profile = ""
|
||||||
# #shared_credential_file = ""
|
# #shared_credential_file = ""
|
||||||
#
|
#
|
||||||
|
# ## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
# ## determined and this option should only be set if you wish to override the
|
||||||
|
# ## default.
|
||||||
|
# ## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# # endpoint_url = ""
|
||||||
|
#
|
||||||
# ## Namespace for the CloudWatch MetricDatums
|
# ## Namespace for the CloudWatch MetricDatums
|
||||||
# namespace = "InfluxData/Telegraf"
|
# namespace = "InfluxData/Telegraf"
|
||||||
|
#
|
||||||
|
# ## If you have a large amount of metrics, you should consider to send statistic
|
||||||
|
# ## values instead of raw metrics which could not only improve performance but
|
||||||
|
# ## also save AWS API cost. If enable this flag, this plugin would parse the required
|
||||||
|
# ## CloudWatch statistic fields (count, min, max, and sum) and send them to CloudWatch.
|
||||||
|
# ## You could use basicstats aggregator to calculate those fields. If not all statistic
|
||||||
|
# ## fields are available, all fields would still be sent as raw metrics.
|
||||||
|
# # write_statistics = false
|
||||||
|
|
||||||
|
|
||||||
# # Configuration for CrateDB to send metrics to.
|
# # Configuration for CrateDB to send metrics to.
|
||||||
|
@ -429,6 +469,12 @@
|
||||||
# # username = "username"
|
# # username = "username"
|
||||||
# # password = "pa$$word"
|
# # password = "pa$$word"
|
||||||
#
|
#
|
||||||
|
# ## OAuth2 Client Credentials Grant
|
||||||
|
# # client_id = "clientid"
|
||||||
|
# # client_secret = "secret"
|
||||||
|
# # token_url = "https://indentityprovider/oauth2/v1/token"
|
||||||
|
# # scopes = ["urn:opc:idm:__myscopes__"]
|
||||||
|
#
|
||||||
# ## Optional TLS Config
|
# ## Optional TLS Config
|
||||||
# # tls_ca = "/etc/telegraf/ca.pem"
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
# # tls_cert = "/etc/telegraf/cert.pem"
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
@ -470,6 +516,15 @@
|
||||||
# ## Kafka topic for producer messages
|
# ## Kafka topic for producer messages
|
||||||
# topic = "telegraf"
|
# topic = "telegraf"
|
||||||
#
|
#
|
||||||
|
# ## Optional Client id
|
||||||
|
# # client_id = "Telegraf"
|
||||||
|
#
|
||||||
|
# ## Set the minimal supported Kafka version. Setting this enables the use of new
|
||||||
|
# ## Kafka features and APIs. Of particular interest, lz4 compression
|
||||||
|
# ## requires at least version 0.10.0.0.
|
||||||
|
# ## ex: version = "1.1.0"
|
||||||
|
# # version = ""
|
||||||
|
#
|
||||||
# ## Optional topic suffix configuration.
|
# ## Optional topic suffix configuration.
|
||||||
# ## If the section is omitted, no suffix is used.
|
# ## If the section is omitted, no suffix is used.
|
||||||
# ## Following topic suffix methods are supported:
|
# ## Following topic suffix methods are supported:
|
||||||
|
@ -501,11 +556,19 @@
|
||||||
# ## ie, if this tag exists, its value will be used as the routing key
|
# ## ie, if this tag exists, its value will be used as the routing key
|
||||||
# routing_tag = "host"
|
# routing_tag = "host"
|
||||||
#
|
#
|
||||||
|
# ## Static routing key. Used when no routing_tag is set or as a fallback
|
||||||
|
# ## when the tag specified in routing tag is not found. If set to "random",
|
||||||
|
# ## a random value will be generated for each message.
|
||||||
|
# ## ex: routing_key = "random"
|
||||||
|
# ## routing_key = "telegraf"
|
||||||
|
# # routing_key = ""
|
||||||
|
#
|
||||||
# ## CompressionCodec represents the various compression codecs recognized by
|
# ## CompressionCodec represents the various compression codecs recognized by
|
||||||
# ## Kafka in messages.
|
# ## Kafka in messages.
|
||||||
# ## 0 : No compression
|
# ## 0 : No compression
|
||||||
# ## 1 : Gzip compression
|
# ## 1 : Gzip compression
|
||||||
# ## 2 : Snappy compression
|
# ## 2 : Snappy compression
|
||||||
|
# ## 3 : LZ4 compression
|
||||||
# # compression_codec = 0
|
# # compression_codec = 0
|
||||||
#
|
#
|
||||||
# ## RequiredAcks is used in Produce Requests to tell the broker how many
|
# ## RequiredAcks is used in Produce Requests to tell the broker how many
|
||||||
|
@ -528,6 +591,10 @@
|
||||||
# ## until the next flush.
|
# ## until the next flush.
|
||||||
# # max_retry = 3
|
# # max_retry = 3
|
||||||
#
|
#
|
||||||
|
# ## The maximum permitted size of a message. Should be set equal to or
|
||||||
|
# ## smaller than the broker's 'message.max.bytes'.
|
||||||
|
# # max_message_bytes = 1000000
|
||||||
|
#
|
||||||
# ## Optional TLS Config
|
# ## Optional TLS Config
|
||||||
# # tls_ca = "/etc/telegraf/ca.pem"
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
# # tls_cert = "/etc/telegraf/cert.pem"
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
@ -566,6 +633,12 @@
|
||||||
# #profile = ""
|
# #profile = ""
|
||||||
# #shared_credential_file = ""
|
# #shared_credential_file = ""
|
||||||
#
|
#
|
||||||
|
# ## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
# ## determined and this option should only be set if you wish to override the
|
||||||
|
# ## default.
|
||||||
|
# ## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# # endpoint_url = ""
|
||||||
|
#
|
||||||
# ## Kinesis StreamName must exist prior to starting telegraf.
|
# ## Kinesis StreamName must exist prior to starting telegraf.
|
||||||
# streamname = "StreamName"
|
# streamname = "StreamName"
|
||||||
# ## DEPRECATED: PartitionKey as used for sharding data.
|
# ## DEPRECATED: PartitionKey as used for sharding data.
|
||||||
|
@ -721,11 +794,11 @@
|
||||||
#
|
#
|
||||||
# ## Number of data points to send to OpenTSDB in Http requests.
|
# ## Number of data points to send to OpenTSDB in Http requests.
|
||||||
# ## Not used with telnet API.
|
# ## Not used with telnet API.
|
||||||
# httpBatchSize = 50
|
# http_batch_size = 50
|
||||||
#
|
#
|
||||||
# ## URI Path for Http requests to OpenTSDB.
|
# ## URI Path for Http requests to OpenTSDB.
|
||||||
# ## Used in cases where OpenTSDB is located behind a reverse proxy.
|
# ## Used in cases where OpenTSDB is located behind a reverse proxy.
|
||||||
# httpPath = "/api/put"
|
# http_path = "/api/put"
|
||||||
#
|
#
|
||||||
# ## Debug true - Prints OpenTSDB communication
|
# ## Debug true - Prints OpenTSDB communication
|
||||||
# debug = false
|
# debug = false
|
||||||
|
@ -737,29 +810,33 @@
|
||||||
# # Configuration for the Prometheus client to spawn
|
# # Configuration for the Prometheus client to spawn
|
||||||
# [[outputs.prometheus_client]]
|
# [[outputs.prometheus_client]]
|
||||||
# ## Address to listen on
|
# ## Address to listen on
|
||||||
# # listen = ":9273"
|
# listen = ":9273"
|
||||||
#
|
#
|
||||||
# ## Use TLS
|
# ## Use HTTP Basic Authentication.
|
||||||
# #tls_cert = "/etc/ssl/telegraf.crt"
|
# # basic_username = "Foo"
|
||||||
# #tls_key = "/etc/ssl/telegraf.key"
|
# # basic_password = "Bar"
|
||||||
#
|
#
|
||||||
# ## Use http basic authentication
|
# ## If set, the IP Ranges which are allowed to access metrics.
|
||||||
# #basic_username = "Foo"
|
# ## ex: ip_range = ["192.168.0.0/24", "192.168.1.0/30"]
|
||||||
# #basic_password = "Bar"
|
# # ip_range = []
|
||||||
#
|
#
|
||||||
# ## IP Ranges which are allowed to access metrics
|
# ## Path to publish the metrics on.
|
||||||
# #ip_range = ["192.168.0.0/24", "192.168.1.0/30"]
|
# # path = "/metrics"
|
||||||
#
|
#
|
||||||
# ## Interval to expire metrics and not deliver to prometheus, 0 == no expiration
|
# ## Expiration interval for each metric. 0 == no expiration
|
||||||
# # expiration_interval = "60s"
|
# # expiration_interval = "60s"
|
||||||
#
|
#
|
||||||
# ## Collectors to enable, valid entries are "gocollector" and "process".
|
# ## Collectors to enable, valid entries are "gocollector" and "process".
|
||||||
# ## If unset, both are enabled.
|
# ## If unset, both are enabled.
|
||||||
# collectors_exclude = ["gocollector", "process"]
|
# # collectors_exclude = ["gocollector", "process"]
|
||||||
#
|
#
|
||||||
# # Send string metrics as Prometheus labels.
|
# ## Send string metrics as Prometheus labels.
|
||||||
# # Unless set to false all string metrics will be sent as labels.
|
# ## Unless set to false all string metrics will be sent as labels.
|
||||||
# string_as_label = true
|
# # string_as_label = true
|
||||||
|
#
|
||||||
|
# ## If set, enable TLS with the given certificate.
|
||||||
|
# # tls_cert = "/etc/ssl/telegraf.crt"
|
||||||
|
# # tls_key = "/etc/ssl/telegraf.key"
|
||||||
|
|
||||||
|
|
||||||
# # Configuration for the Riemann server to send metrics to
|
# # Configuration for the Riemann server to send metrics to
|
||||||
|
@ -913,6 +990,27 @@
|
||||||
# float = []
|
# float = []
|
||||||
|
|
||||||
|
|
||||||
|
# # Map enum values according to given table.
|
||||||
|
# [[processors.enum]]
|
||||||
|
# [[processors.enum.fields]]
|
||||||
|
# ## Name of the field to map
|
||||||
|
# source = "name"
|
||||||
|
#
|
||||||
|
# ## Destination field to be used for the mapped value. By default the source
|
||||||
|
# ## field is used, overwriting the original value.
|
||||||
|
# # destination = "mapped"
|
||||||
|
#
|
||||||
|
# ## Default value to be used for all values not contained in the mapping
|
||||||
|
# ## table. When unset, the unmodified value for the field will be used if no
|
||||||
|
# ## match is found.
|
||||||
|
# # default = 0
|
||||||
|
#
|
||||||
|
# ## Table of mappings
|
||||||
|
# [processors.enum.fields.value_mappings]
|
||||||
|
# value1 = 1
|
||||||
|
# value2 = 2
|
||||||
|
|
||||||
|
|
||||||
# # Apply metric modifications using override semantics.
|
# # Apply metric modifications using override semantics.
|
||||||
# [[processors.override]]
|
# [[processors.override]]
|
||||||
# ## All modifications on inputs and aggregators can be overridden:
|
# ## All modifications on inputs and aggregators can be overridden:
|
||||||
|
@ -925,6 +1023,25 @@
|
||||||
# # additional_tag = "tag_value"
|
# # additional_tag = "tag_value"
|
||||||
|
|
||||||
|
|
||||||
|
# # Parse a value in a specified field/tag(s) and add the result in a new metric
|
||||||
|
# [[processors.parser]]
|
||||||
|
# ## The name of the fields whose value will be parsed.
|
||||||
|
# parse_fields = []
|
||||||
|
#
|
||||||
|
# ## If true, incoming metrics are not emitted.
|
||||||
|
# drop_original = false
|
||||||
|
#
|
||||||
|
# ## If set to override, emitted metrics will be merged by overriding the
|
||||||
|
# ## original metric using the newly parsed metrics.
|
||||||
|
# merge = "override"
|
||||||
|
#
|
||||||
|
# ## The dataformat to be read from files
|
||||||
|
# ## Each data format has its own unique set of configuration options, read
|
||||||
|
# ## more about them here:
|
||||||
|
# ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
||||||
|
|
||||||
# # Print all metrics that pass through this filter.
|
# # Print all metrics that pass through this filter.
|
||||||
# [[processors.printer]]
|
# [[processors.printer]]
|
||||||
|
|
||||||
|
@ -959,6 +1076,67 @@
|
||||||
# # result_key = "search_category"
|
# # result_key = "search_category"
|
||||||
|
|
||||||
|
|
||||||
|
# # Rename measurements, tags, and fields that pass through this filter.
|
||||||
|
# [[processors.rename]]
|
||||||
|
# ## Measurement, tag, and field renamings are stored in separate sub-tables.
|
||||||
|
# ## Specify one sub-table per rename operation.
|
||||||
|
# # [[processors.rename.measurement]]
|
||||||
|
# # ## measurement to change
|
||||||
|
# # from = "kilobytes_per_second"
|
||||||
|
# # to = "kbps"
|
||||||
|
#
|
||||||
|
# # [[processors.rename.tag]]
|
||||||
|
# # ## tag to change
|
||||||
|
# # from = "host"
|
||||||
|
# # to = "hostname"
|
||||||
|
#
|
||||||
|
# # [[processors.rename.field]]
|
||||||
|
# # ## field to change
|
||||||
|
# # from = "lower"
|
||||||
|
# # to = "min"
|
||||||
|
#
|
||||||
|
# # [[processors.rename.field]]
|
||||||
|
# # ## field to change
|
||||||
|
# # from = "upper"
|
||||||
|
# # to = "max"
|
||||||
|
|
||||||
|
|
||||||
|
# # Perform string processing on tags, fields, and measurements
|
||||||
|
# [[processors.strings]]
|
||||||
|
# ## Convert a tag value to uppercase
|
||||||
|
# # [[processors.strings.uppercase]]
|
||||||
|
# # tag = "method"
|
||||||
|
#
|
||||||
|
# ## Convert a field value to lowercase and store in a new field
|
||||||
|
# # [[processors.strings.lowercase]]
|
||||||
|
# # field = "uri_stem"
|
||||||
|
# # dest = "uri_stem_normalised"
|
||||||
|
#
|
||||||
|
# ## Trim leading and trailing whitespace using the default cutset
|
||||||
|
# # [[processors.strings.trim]]
|
||||||
|
# # field = "message"
|
||||||
|
#
|
||||||
|
# ## Trim leading characters in cutset
|
||||||
|
# # [[processors.strings.trim_left]]
|
||||||
|
# # field = "message"
|
||||||
|
# # cutset = "\t"
|
||||||
|
#
|
||||||
|
# ## Trim trailing characters in cutset
|
||||||
|
# # [[processors.strings.trim_right]]
|
||||||
|
# # field = "message"
|
||||||
|
# # cutset = "\r\n"
|
||||||
|
#
|
||||||
|
# ## Trim the given prefix from the field
|
||||||
|
# # [[processors.strings.trim_prefix]]
|
||||||
|
# # field = "my_value"
|
||||||
|
# # prefix = "my_"
|
||||||
|
#
|
||||||
|
# ## Trim the given suffix from the field
|
||||||
|
# # [[processors.strings.trim_suffix]]
|
||||||
|
# # field = "read_count"
|
||||||
|
# # suffix = "_count"
|
||||||
|
|
||||||
|
|
||||||
# # Print all metrics that pass through this filter.
|
# # Print all metrics that pass through this filter.
|
||||||
# [[processors.topk]]
|
# [[processors.topk]]
|
||||||
# ## How many seconds between aggregations
|
# ## How many seconds between aggregations
|
||||||
|
@ -1060,6 +1238,18 @@
|
||||||
# drop_original = false
|
# drop_original = false
|
||||||
|
|
||||||
|
|
||||||
|
# # Count the occurance of values in fields.
|
||||||
|
# [[aggregators.valuecounter]]
|
||||||
|
# ## General Aggregator Arguments:
|
||||||
|
# ## The period on which to flush & clear the aggregator.
|
||||||
|
# period = "30s"
|
||||||
|
# ## If true, the original metric will be dropped by the
|
||||||
|
# ## aggregator and will not get sent to the output plugins.
|
||||||
|
# drop_original = false
|
||||||
|
# ## The fields for which the values will be counted
|
||||||
|
# fields = []
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
###############################################################################
|
###############################################################################
|
||||||
# INPUT PLUGINS #
|
# INPUT PLUGINS #
|
||||||
|
@ -1139,6 +1329,31 @@
|
||||||
# no configuration
|
# no configuration
|
||||||
|
|
||||||
|
|
||||||
|
# # Gather ActiveMQ metrics
|
||||||
|
# [[inputs.activemq]]
|
||||||
|
# ## Required ActiveMQ Endpoint
|
||||||
|
# # server = "192.168.50.10"
|
||||||
|
#
|
||||||
|
# ## Required ActiveMQ port
|
||||||
|
# # port = 8161
|
||||||
|
#
|
||||||
|
# ## Credentials for basic HTTP authentication
|
||||||
|
# # username = "admin"
|
||||||
|
# # password = "admin"
|
||||||
|
#
|
||||||
|
# ## Required ActiveMQ webadmin root path
|
||||||
|
# # webadmin = "admin"
|
||||||
|
#
|
||||||
|
# ## Maximum time to receive response.
|
||||||
|
# # response_timeout = "5s"
|
||||||
|
#
|
||||||
|
# ## Optional TLS Config
|
||||||
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# # tls_key = "/etc/telegraf/key.pem"
|
||||||
|
# ## Use TLS but skip chain & host verification
|
||||||
|
|
||||||
|
|
||||||
# # Read stats from aerospike server(s)
|
# # Read stats from aerospike server(s)
|
||||||
# [[inputs.aerospike]]
|
# [[inputs.aerospike]]
|
||||||
# ## Aerospike servers to connect to (with port)
|
# ## Aerospike servers to connect to (with port)
|
||||||
|
@ -1349,6 +1564,12 @@
|
||||||
# #profile = ""
|
# #profile = ""
|
||||||
# #shared_credential_file = ""
|
# #shared_credential_file = ""
|
||||||
#
|
#
|
||||||
|
# ## Endpoint to make request against, the correct endpoint is automatically
|
||||||
|
# ## determined and this option should only be set if you wish to override the
|
||||||
|
# ## default.
|
||||||
|
# ## ex: endpoint_url = "http://localhost:8000"
|
||||||
|
# # endpoint_url = ""
|
||||||
|
#
|
||||||
# # The minimum period for Cloudwatch metrics is 1 minute (60s). However not all
|
# # The minimum period for Cloudwatch metrics is 1 minute (60s). However not all
|
||||||
# # metrics are made available to the 1 minute period. Some are collected at
|
# # metrics are made available to the 1 minute period. Some are collected at
|
||||||
# # 3 minute, 5 minute, or larger intervals. See https://aws.amazon.com/cloudwatch/faqs/#monitoring.
|
# # 3 minute, 5 minute, or larger intervals. See https://aws.amazon.com/cloudwatch/faqs/#monitoring.
|
||||||
|
@ -1385,7 +1606,9 @@
|
||||||
# #[[inputs.cloudwatch.metrics]]
|
# #[[inputs.cloudwatch.metrics]]
|
||||||
# # names = ["Latency", "RequestCount"]
|
# # names = ["Latency", "RequestCount"]
|
||||||
# #
|
# #
|
||||||
# # ## Dimension filters for Metric (optional)
|
# # ## Dimension filters for Metric. These are optional however all dimensions
|
||||||
|
# # ## defined for the metric names must be specified in order to retrieve
|
||||||
|
# # ## the metric statistics.
|
||||||
# # [[inputs.cloudwatch.metrics.dimensions]]
|
# # [[inputs.cloudwatch.metrics.dimensions]]
|
||||||
# # name = "LoadBalancerName"
|
# # name = "LoadBalancerName"
|
||||||
# # value = "p-example"
|
# # value = "p-example"
|
||||||
|
@ -1455,7 +1678,7 @@
|
||||||
# # Read CouchDB Stats from one or more servers
|
# # Read CouchDB Stats from one or more servers
|
||||||
# [[inputs.couchdb]]
|
# [[inputs.couchdb]]
|
||||||
# ## Works with CouchDB stats endpoints out of the box
|
# ## Works with CouchDB stats endpoints out of the box
|
||||||
# ## Multiple HOSTs from which to read CouchDB stats:
|
# ## Multiple Hosts from which to read CouchDB stats:
|
||||||
# hosts = ["http://localhost:8086/_stats"]
|
# hosts = ["http://localhost:8086/_stats"]
|
||||||
|
|
||||||
|
|
||||||
|
@ -1685,6 +1908,48 @@
|
||||||
# # timeout = "5s"
|
# # timeout = "5s"
|
||||||
|
|
||||||
|
|
||||||
|
# # Reload and gather from file[s] on telegraf's interval.
|
||||||
|
# [[inputs.file]]
|
||||||
|
# ## Files to parse each interval.
|
||||||
|
# ## These accept standard unix glob matching rules, but with the addition of
|
||||||
|
# ## ** as a "super asterisk". ie:
|
||||||
|
# ## /var/log/**.log -> recursively find all .log files in /var/log
|
||||||
|
# ## /var/log/*/*.log -> find all .log files with a parent dir in /var/log
|
||||||
|
# ## /var/log/apache.log -> only read the apache log file
|
||||||
|
# files = ["/var/log/apache/access.log"]
|
||||||
|
#
|
||||||
|
# ## The dataformat to be read from files
|
||||||
|
# ## Each data format has its own unique set of configuration options, read
|
||||||
|
# ## more about them here:
|
||||||
|
# ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
|
||||||
|
# data_format = "influx"
|
||||||
|
|
||||||
|
|
||||||
|
# # Count files in a directory
|
||||||
|
# [[inputs.filecount]]
|
||||||
|
# ## Directory to gather stats about.
|
||||||
|
# directory = "/var/cache/apt/archives"
|
||||||
|
#
|
||||||
|
# ## Only count files that match the name pattern. Defaults to "*".
|
||||||
|
# name = "*.deb"
|
||||||
|
#
|
||||||
|
# ## Count files in subdirectories. Defaults to true.
|
||||||
|
# recursive = false
|
||||||
|
#
|
||||||
|
# ## Only count regular files. Defaults to true.
|
||||||
|
# regular_only = true
|
||||||
|
#
|
||||||
|
# ## Only count files that are at least this size in bytes. If size is
|
||||||
|
# ## a negative number, only count files that are smaller than the
|
||||||
|
# ## absolute value of size. Defaults to 0.
|
||||||
|
# size = 0
|
||||||
|
#
|
||||||
|
# ## Only count files that have not been touched for at least this
|
||||||
|
# ## duration. If mtime is negative, only count files that have been
|
||||||
|
# ## touched in this duration. Defaults to "0s".
|
||||||
|
# mtime = "0s"
|
||||||
|
|
||||||
|
|
||||||
# # Read stats about given file(s)
|
# # Read stats about given file(s)
|
||||||
# [[inputs.filestat]]
|
# [[inputs.filestat]]
|
||||||
# ## Files to gather stats about.
|
# ## Files to gather stats about.
|
||||||
|
@ -1919,6 +2184,29 @@
|
||||||
# # apiVersion = "v1"
|
# # apiVersion = "v1"
|
||||||
|
|
||||||
|
|
||||||
|
# # Gather Icinga2 status
|
||||||
|
# [[inputs.icinga2]]
|
||||||
|
# ## Required Icinga2 server address (default: "https://localhost:5665")
|
||||||
|
# # server = "https://localhost:5665"
|
||||||
|
#
|
||||||
|
# ## Required Icinga2 object type ("services" or "hosts, default "services")
|
||||||
|
# # object_type = "services"
|
||||||
|
#
|
||||||
|
# ## Credentials for basic HTTP authentication
|
||||||
|
# # username = "admin"
|
||||||
|
# # password = "admin"
|
||||||
|
#
|
||||||
|
# ## Maximum time to receive response.
|
||||||
|
# # response_timeout = "5s"
|
||||||
|
#
|
||||||
|
# ## Optional TLS Config
|
||||||
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# # tls_key = "/etc/telegraf/key.pem"
|
||||||
|
# ## Use TLS but skip chain & host verification
|
||||||
|
# # insecure_skip_verify = true
|
||||||
|
|
||||||
|
|
||||||
# # Read InfluxDB-formatted JSON metrics from one or more HTTP endpoints
|
# # Read InfluxDB-formatted JSON metrics from one or more HTTP endpoints
|
||||||
# [[inputs.influxdb]]
|
# [[inputs.influxdb]]
|
||||||
# ## Works with InfluxDB debug endpoints out of the box,
|
# ## Works with InfluxDB debug endpoints out of the box,
|
||||||
|
@ -1978,7 +2266,7 @@
|
||||||
#
|
#
|
||||||
# ## Timeout for the ipmitool command to complete
|
# ## Timeout for the ipmitool command to complete
|
||||||
# timeout = "20s"
|
# timeout = "20s"
|
||||||
|
#
|
||||||
# ## Schema Version: (Optional, defaults to version 1)
|
# ## Schema Version: (Optional, defaults to version 1)
|
||||||
# metric_version = 2
|
# metric_version = 2
|
||||||
|
|
||||||
|
@ -2004,8 +2292,6 @@
|
||||||
# ## Setting 'use_lock' to true runs iptables with the "-w" option.
|
# ## Setting 'use_lock' to true runs iptables with the "-w" option.
|
||||||
# ## Adjust your sudo settings appropriately if using this option ("iptables -wnvl")
|
# ## Adjust your sudo settings appropriately if using this option ("iptables -wnvl")
|
||||||
# use_lock = false
|
# use_lock = false
|
||||||
# ## Define an alternate executable, such as "ip6tables". Default is "iptables".
|
|
||||||
# # binary = "ip6tables"
|
|
||||||
# ## defines the table to monitor:
|
# ## defines the table to monitor:
|
||||||
# table = "filter"
|
# table = "filter"
|
||||||
# ## defines the chains to monitor.
|
# ## defines the chains to monitor.
|
||||||
|
@ -2164,6 +2450,26 @@
|
||||||
# # no configuration
|
# # no configuration
|
||||||
|
|
||||||
|
|
||||||
|
# # Read status information from one or more Kibana servers
|
||||||
|
# [[inputs.kibana]]
|
||||||
|
# ## specify a list of one or more Kibana servers
|
||||||
|
# servers = ["http://localhost:5601"]
|
||||||
|
#
|
||||||
|
# ## Timeout for HTTP requests
|
||||||
|
# timeout = "5s"
|
||||||
|
#
|
||||||
|
# ## HTTP Basic Auth credentials
|
||||||
|
# # username = "username"
|
||||||
|
# # password = "pa$$word"
|
||||||
|
#
|
||||||
|
# ## Optional TLS Config
|
||||||
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# # tls_key = "/etc/telegraf/key.pem"
|
||||||
|
# ## Use TLS but skip chain & host verification
|
||||||
|
# # insecure_skip_verify = false
|
||||||
|
|
||||||
|
|
||||||
# # Read metrics from the kubernetes kubelet api
|
# # Read metrics from the kubernetes kubelet api
|
||||||
# [[inputs.kubernetes]]
|
# [[inputs.kubernetes]]
|
||||||
# ## URL for the kubelet
|
# ## URL for the kubelet
|
||||||
|
@ -2348,7 +2654,7 @@
|
||||||
# ## gather thread state counts from INFORMATION_SCHEMA.PROCESSLIST
|
# ## gather thread state counts from INFORMATION_SCHEMA.PROCESSLIST
|
||||||
# gather_process_list = true
|
# gather_process_list = true
|
||||||
# #
|
# #
|
||||||
# ## gather thread state counts from INFORMATION_SCHEMA.USER_STATISTICS
|
# ## gather user statistics from INFORMATION_SCHEMA.USER_STATISTICS
|
||||||
# gather_user_statistics = true
|
# gather_user_statistics = true
|
||||||
# #
|
# #
|
||||||
# ## gather auto_increment columns and max values from information schema
|
# ## gather auto_increment columns and max values from information schema
|
||||||
|
@ -2499,11 +2805,11 @@
|
||||||
|
|
||||||
# # Pulls statistics from nvidia GPUs attached to the host
|
# # Pulls statistics from nvidia GPUs attached to the host
|
||||||
# [[inputs.nvidia_smi]]
|
# [[inputs.nvidia_smi]]
|
||||||
# ## Optional: path to nvidia-smi binary, defaults to $PATH via exec.LookPath
|
# ## Optional: path to nvidia-smi binary, defaults to $PATH via exec.LookPath
|
||||||
# # bin_path = /usr/bin/nvidia-smi
|
# # bin_path = "/usr/bin/nvidia-smi"
|
||||||
#
|
#
|
||||||
# ## Optional: timeout for GPU polling
|
# ## Optional: timeout for GPU polling
|
||||||
# # timeout = 5s
|
# # timeout = "5s"
|
||||||
|
|
||||||
|
|
||||||
# # OpenLDAP cn=Monitor plugin
|
# # OpenLDAP cn=Monitor plugin
|
||||||
|
@ -2514,7 +2820,7 @@
|
||||||
# # ldaps, starttls, or no encryption. default is an empty string, disabling all encryption.
|
# # ldaps, starttls, or no encryption. default is an empty string, disabling all encryption.
|
||||||
# # note that port will likely need to be changed to 636 for ldaps
|
# # note that port will likely need to be changed to 636 for ldaps
|
||||||
# # valid options: "" | "starttls" | "ldaps"
|
# # valid options: "" | "starttls" | "ldaps"
|
||||||
# ssl = ""
|
# tls = ""
|
||||||
#
|
#
|
||||||
# # skip peer certificate verification. Default is false.
|
# # skip peer certificate verification. Default is false.
|
||||||
# insecure_skip_verify = false
|
# insecure_skip_verify = false
|
||||||
|
@ -2591,20 +2897,23 @@
|
||||||
|
|
||||||
# # Ping given url(s) and return statistics
|
# # Ping given url(s) and return statistics
|
||||||
# [[inputs.ping]]
|
# [[inputs.ping]]
|
||||||
# ## NOTE: this plugin forks the ping command. You may need to set capabilities
|
|
||||||
# ## via setcap cap_net_raw+p /bin/ping
|
|
||||||
# #
|
|
||||||
# ## List of urls to ping
|
# ## List of urls to ping
|
||||||
# urls = ["www.google.com"] # required
|
# urls = ["example.org"]
|
||||||
# ## number of pings to send per collection (ping -c <COUNT>)
|
#
|
||||||
|
# ## Number of pings to send per collection (ping -c <COUNT>)
|
||||||
# # count = 1
|
# # count = 1
|
||||||
# ## interval, in s, at which to ping. 0 == default (ping -i <PING_INTERVAL>)
|
#
|
||||||
|
# ## Interval, in s, at which to ping. 0 == default (ping -i <PING_INTERVAL>)
|
||||||
|
# ## Not available in Windows.
|
||||||
# # ping_interval = 1.0
|
# # ping_interval = 1.0
|
||||||
# ## per-ping timeout, in s. 0 == no timeout (ping -W <TIMEOUT>)
|
#
|
||||||
|
# ## Per-ping timeout, in s. 0 == no timeout (ping -W <TIMEOUT>)
|
||||||
# # timeout = 1.0
|
# # timeout = 1.0
|
||||||
# ## total-ping deadline, in s. 0 == no deadline (ping -w <DEADLINE>)
|
#
|
||||||
|
# ## Total-ping deadline, in s. 0 == no deadline (ping -w <DEADLINE>)
|
||||||
# # deadline = 10
|
# # deadline = 10
|
||||||
# ## interface or source address to send ping from (ping -I <INTERFACE/SRC_ADDR>)
|
#
|
||||||
|
# ## Interface or source address to send ping from (ping -I <INTERFACE/SRC_ADDR>)
|
||||||
# ## on Darwin and Freebsd only source address possible: (ping -S <SRC_ADDR>)
|
# ## on Darwin and Freebsd only source address possible: (ping -S <SRC_ADDR>)
|
||||||
# # interface = ""
|
# # interface = ""
|
||||||
|
|
||||||
|
@ -2748,6 +3057,13 @@
|
||||||
# ## If no servers are specified, then localhost is used as the host.
|
# ## If no servers are specified, then localhost is used as the host.
|
||||||
# ## If no port is specified, 6379 is used
|
# ## If no port is specified, 6379 is used
|
||||||
# servers = ["tcp://localhost:6379"]
|
# servers = ["tcp://localhost:6379"]
|
||||||
|
#
|
||||||
|
# ## Optional TLS Config
|
||||||
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# # tls_key = "/etc/telegraf/key.pem"
|
||||||
|
# ## Use TLS but skip chain & host verification
|
||||||
|
# # insecure_skip_verify = true
|
||||||
|
|
||||||
|
|
||||||
# # Read metrics from one or many RethinkDB servers
|
# # Read metrics from one or many RethinkDB servers
|
||||||
|
@ -3119,6 +3435,22 @@
|
||||||
# # virtual_servers = [1]
|
# # virtual_servers = [1]
|
||||||
|
|
||||||
|
|
||||||
|
# # Read Tengine's basic status information (ngx_http_reqstat_module)
|
||||||
|
# [[inputs.tengine]]
|
||||||
|
# # An array of Tengine reqstat module URI to gather stats.
|
||||||
|
# urls = ["http://127.0.0.1/us"]
|
||||||
|
#
|
||||||
|
# # HTTP response timeout (default: 5s)
|
||||||
|
# # response_timeout = "5s"
|
||||||
|
#
|
||||||
|
# ## Optional TLS Config
|
||||||
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# # tls_cert = "/etc/telegraf/cert.cer"
|
||||||
|
# # tls_key = "/etc/telegraf/key.key"
|
||||||
|
# ## Use TLS but skip chain & host verification
|
||||||
|
# # insecure_skip_verify = false
|
||||||
|
|
||||||
|
|
||||||
# # Gather metrics from the Tomcat server status page.
|
# # Gather metrics from the Tomcat server status page.
|
||||||
# [[inputs.tomcat]]
|
# [[inputs.tomcat]]
|
||||||
# ## URL of the Tomcat server status
|
# ## URL of the Tomcat server status
|
||||||
|
@ -3195,6 +3527,23 @@
|
||||||
# # instance_name = instanceName
|
# # instance_name = instanceName
|
||||||
|
|
||||||
|
|
||||||
|
# # Reads metrics from a SSL certificate
|
||||||
|
# [[inputs.x509_cert]]
|
||||||
|
# ## List certificate sources
|
||||||
|
# sources = ["/etc/ssl/certs/ssl-cert-snakeoil.pem", "tcp://example.org:443"]
|
||||||
|
#
|
||||||
|
# ## Timeout for SSL connection
|
||||||
|
# # timeout = "5s"
|
||||||
|
#
|
||||||
|
# ## Optional TLS Config
|
||||||
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
# # tls_key = "/etc/telegraf/key.pem"
|
||||||
|
#
|
||||||
|
# ## Use TLS but skip chain & host verification
|
||||||
|
# # insecure_skip_verify = false
|
||||||
|
|
||||||
|
|
||||||
# # Read metrics of ZFS from arcstats, zfetchstats, vdev_cache_stats, and pools
|
# # Read metrics of ZFS from arcstats, zfetchstats, vdev_cache_stats, and pools
|
||||||
# [[inputs.zfs]]
|
# [[inputs.zfs]]
|
||||||
# ## ZFS kstat path. Ignored on FreeBSD
|
# ## ZFS kstat path. Ignored on FreeBSD
|
||||||
|
@ -3268,10 +3617,13 @@
|
||||||
# # exchange_arguments = { }
|
# # exchange_arguments = { }
|
||||||
# # exchange_arguments = {"hash_propery" = "timestamp"}
|
# # exchange_arguments = {"hash_propery" = "timestamp"}
|
||||||
#
|
#
|
||||||
# ## AMQP queue name
|
# ## AMQP queue name.
|
||||||
# queue = "telegraf"
|
# queue = "telegraf"
|
||||||
#
|
#
|
||||||
# ## Binding Key
|
# ## AMQP queue durability can be "transient" or "durable".
|
||||||
|
# queue_durability = "durable"
|
||||||
|
#
|
||||||
|
# ## Binding Key.
|
||||||
# binding_key = "#"
|
# binding_key = "#"
|
||||||
#
|
#
|
||||||
# ## Maximum number of messages server should give to the worker.
|
# ## Maximum number of messages server should give to the worker.
|
||||||
|
@ -3404,6 +3756,15 @@
|
||||||
# ## topic(s) to consume
|
# ## topic(s) to consume
|
||||||
# topics = ["telegraf"]
|
# topics = ["telegraf"]
|
||||||
#
|
#
|
||||||
|
# ## Optional Client id
|
||||||
|
# # client_id = "Telegraf"
|
||||||
|
#
|
||||||
|
# ## Set the minimal supported Kafka version. Setting this enables the use of new
|
||||||
|
# ## Kafka features and APIs. Of particular interest, lz4 compression
|
||||||
|
# ## requires at least version 0.10.0.0.
|
||||||
|
# ## ex: version = "1.1.0"
|
||||||
|
# # version = ""
|
||||||
|
#
|
||||||
# ## Optional TLS Config
|
# ## Optional TLS Config
|
||||||
# # tls_ca = "/etc/telegraf/ca.pem"
|
# # tls_ca = "/etc/telegraf/ca.pem"
|
||||||
# # tls_cert = "/etc/telegraf/cert.pem"
|
# # tls_cert = "/etc/telegraf/cert.pem"
|
||||||
|
@ -3428,7 +3789,7 @@
|
||||||
#
|
#
|
||||||
# ## Maximum length of a message to consume, in bytes (default 0/unlimited);
|
# ## Maximum length of a message to consume, in bytes (default 0/unlimited);
|
||||||
# ## larger messages are dropped
|
# ## larger messages are dropped
|
||||||
# max_message_len = 65536
|
# max_message_len = 1000000
|
||||||
|
|
||||||
|
|
||||||
# # Read metrics from Kafka topic(s)
|
# # Read metrics from Kafka topic(s)
|
||||||
|
@ -3457,6 +3818,9 @@
|
||||||
|
|
||||||
# # Stream and parse log file(s).
|
# # Stream and parse log file(s).
|
||||||
# [[inputs.logparser]]
|
# [[inputs.logparser]]
|
||||||
|
# ## DEPRECATED: The 'logparser' plugin is deprecated in 1.8. Please use the
|
||||||
|
# ## 'tail' plugin with the grok data_format as a replacement.
|
||||||
|
#
|
||||||
# ## Log files to parse.
|
# ## Log files to parse.
|
||||||
# ## These accept standard unix glob matching rules, but with the addition of
|
# ## These accept standard unix glob matching rules, but with the addition of
|
||||||
# ## ** as a "super asterisk". ie:
|
# ## ** as a "super asterisk". ie:
|
||||||
|
@ -3590,6 +3954,19 @@
|
||||||
# data_format = "influx"
|
# data_format = "influx"
|
||||||
|
|
||||||
|
|
||||||
|
# # Read metrics from one or many pgbouncer servers
|
||||||
|
# [[inputs.pgbouncer]]
|
||||||
|
# ## specify address via a url matching:
|
||||||
|
# ## postgres://[pqgotest[:password]]@localhost[/dbname]\
|
||||||
|
# ## ?sslmode=[disable|verify-ca|verify-full]
|
||||||
|
# ## or a simple string:
|
||||||
|
# ## host=localhost user=pqotest password=... sslmode=... dbname=app_production
|
||||||
|
# ##
|
||||||
|
# ## All connection parameters are optional.
|
||||||
|
# ##
|
||||||
|
# address = "host=localhost user=pgbouncer sslmode=disable"
|
||||||
|
|
||||||
|
|
||||||
# # Read metrics from one or many postgresql servers
|
# # Read metrics from one or many postgresql servers
|
||||||
# [[inputs.postgresql]]
|
# [[inputs.postgresql]]
|
||||||
# ## specify address via a url matching:
|
# ## specify address via a url matching:
|
||||||
|
@ -3820,9 +4197,9 @@
|
||||||
# ## Only applies to stream sockets (e.g. TCP).
|
# ## Only applies to stream sockets (e.g. TCP).
|
||||||
# # max_connections = 1024
|
# # max_connections = 1024
|
||||||
#
|
#
|
||||||
# ## Read timeout (default = 500ms).
|
# ## Read timeout is the maximum time allowed for reading a single message (default = 5s).
|
||||||
# ## 0 means unlimited.
|
# ## 0 means unlimited.
|
||||||
# # read_timeout = 500ms
|
# # read_timeout = "5s"
|
||||||
#
|
#
|
||||||
# ## Whether to parse in best effort mode or not (default = false).
|
# ## Whether to parse in best effort mode or not (default = false).
|
||||||
# ## By default best effort parsing is off.
|
# ## By default best effort parsing is off.
|
||||||
|
|
|
@ -95,7 +95,7 @@ supported_packages = {
|
||||||
"freebsd": [ "tar" ]
|
"freebsd": [ "tar" ]
|
||||||
}
|
}
|
||||||
|
|
||||||
next_version = '1.8.0'
|
next_version = '1.9.0'
|
||||||
|
|
||||||
################
|
################
|
||||||
#### Telegraf Functions
|
#### Telegraf Functions
|
||||||
|
|
Loading…
Reference in New Issue