Update etc/telegraf.conf

This commit is contained in:
Daniel Nelson 2020-03-18 13:49:47 -07:00
parent 4e28e5721b
commit e376228a1a
No known key found for this signature in database
GPG Key ID: CAAD59C9444F6155
1 changed files with 442 additions and 45 deletions

View File

@ -131,6 +131,13 @@
## the default retention policy. Only takes effect when using HTTP.
# retention_policy = ""
## The value of this tag will be used to determine the retention policy. If this
## tag is not set the 'retention_policy' option is used as the default.
# retention_policy_tag = ""
## If true, the 'retention_policy_tag' will not be removed from the metric.
# exclude_retention_policy_tag = false
## Write consistency (clusters only), can be: "any", "one", "quorum", "all".
## Only takes effect when using HTTP.
# write_consistency = "any"
@ -577,10 +584,15 @@
# # insecure_skip_verify = false
# # Send telegraf metrics to graylog(s)
# # Send telegraf metrics to graylog
# [[outputs.graylog]]
# ## UDP endpoint for your graylog instance.
# servers = ["127.0.0.1:12201", "192.168.1.1:12201"]
# servers = ["127.0.0.1:12201"]
#
# ## The field to use as the GELF short_message, if unset the static string
# ## "telegraf" will be used.
# ## example: short_message_field = "message"
# # short_message_field = ""
# # Configurable HTTP health check resource based on metrics
@ -742,6 +754,13 @@
# ## Kafka topic for producer messages
# topic = "telegraf"
#
# ## The value of this tag will be used as the topic. If not set the 'topic'
# ## option is used.
# # topic_tag = ""
#
# ## If true, the 'topic_tag' will be removed from to the metric.
# # exclude_topic_tag = false
#
# ## Optional Client id
# # client_id = "Telegraf"
#
@ -778,13 +797,21 @@
# # keys = ["foo", "bar"]
# # separator = "_"
#
# ## Telegraf tag to use as a routing key
# ## ie, if this tag exists, its value will be used as the routing key
# ## The routing tag specifies a tagkey on the metric whose value is used as
# ## the message key. The message key is used to determine which partition to
# ## send the message to. This tag is prefered over the routing_key option.
# routing_tag = "host"
#
# ## Static routing key. Used when no routing_tag is set or as a fallback
# ## when the tag specified in routing tag is not found. If set to "random",
# ## a random value will be generated for each message.
# ## The routing key is set as the message key and used to determine which
# ## partition to send the message to. This value is only used when no
# ## routing_tag is set or as a fallback when the tag specified in routing tag
# ## is not found.
# ##
# ## If set to "random", a random value will be generated for each message.
# ##
# ## When unset, no message key is added and each message is routed to a random
# ## partition.
# ##
# ## ex: routing_key = "random"
# ## routing_key = "telegraf"
# # routing_key = ""
@ -980,9 +1007,14 @@
# [[outputs.nats]]
# ## URLs of NATS servers
# servers = ["nats://localhost:4222"]
#
# ## Optional credentials
# # username = ""
# # password = ""
#
# ## Optional NATS 2.0 and NATS NGS compatible user credentials
# # credentials = "/etc/telegraf/nats.creds"
#
# ## NATS subject for producer messages
# subject = "telegraf"
#
@ -1266,6 +1298,34 @@
# # default_appname = "Telegraf"
# # Write metrics to Warp 10
# [[outputs.warp10]]
# # Prefix to add to the measurement.
# prefix = "telegraf."
#
# # URL of the Warp 10 server
# warp_url = "http://localhost:8080"
#
# # Write token to access your app on warp 10
# token = "Token"
#
# # Warp 10 query timeout
# # timeout = "15s"
#
# ## Print Warp 10 error body
# # print_error_body = false
#
# ## Max string error size
# # max_string_error_size = 511
#
# ## Optional TLS Config
# # tls_ca = "/etc/telegraf/ca.pem"
# # tls_cert = "/etc/telegraf/cert.pem"
# # tls_key = "/etc/telegraf/key.pem"
# ## Use TLS but skip chain & host verification
# # insecure_skip_verify = false
# # Configuration for Wavefront server to send metrics to
# [[outputs.wavefront]]
# ## Url for Wavefront Direct Ingestion or using HTTP with Wavefront Proxy
@ -1341,6 +1401,7 @@
# ## select the keys to convert. The array may contain globs.
# ## <target-type> = [<tag-key>...]
# [processors.converter.tags]
# measurement = []
# string = []
# integer = []
# unsigned = []
@ -1353,6 +1414,7 @@
# ## select the keys to convert. The array may contain globs.
# ## <target-type> = [<field-key>...]
# [processors.converter.fields]
# measurement = []
# tag = []
# string = []
# integer = []
@ -1369,6 +1431,20 @@
# ## Date format string, must be a representation of the Go "reference time"
# ## which is "Mon Jan 2 15:04:05 -0700 MST 2006".
# date_format = "Jan"
#
# ## Offset duration added to the date string when writing the new tag.
# # date_offset = "0s"
#
# ## Timezone to use when creating the tag. This can be set to one of
# ## "UTC", "Local", or to a location name in the IANA Time Zone database.
# ## example: timezone = "America/Los_Angeles"
# # timezone = "UTC"
# # Filter metrics with repeating field values
# [[processors.dedup]]
# ## Maximum time to suppress output
# dedup_interval = "600s"
# # Map enum values according to given table.
@ -1475,6 +1551,20 @@
# [[processors.rename]]
# # Add the S2 Cell ID as a tag based on latitude and longitude fields
# [[processors.s2geo]]
# ## The name of the lat and lon fields containing WGS-84 latitude and
# ## longitude in decimal degrees.
# # lat_field = "lat"
# # lon_field = "lon"
#
# ## New tag to create
# # tag_key = "s2_cell_id"
#
# ## Cell level (see https://s2geometry.io/resources/s2cell_statistics.html)
# # cell_level = 9
# # Perform string processing on tags, fields, and measurements
# [[processors.strings]]
# ## Convert a tag value to uppercase
@ -1486,6 +1576,10 @@
# # field = "uri_stem"
# # dest = "uri_stem_normalised"
#
# ## Convert a field value to titlecase
# # [[processors.strings.titlecase]]
# # field = "status"
#
# ## Trim leading and trailing whitespace using the default cutset
# # [[processors.strings.trim]]
# # field = "message"
@ -1535,6 +1629,17 @@
# keep = ["foo", "bar", "baz"]
# # Uses a Go template to create a new tag
# [[processors.template]]
# ## Tag to set with the output of the template.
# tag = "topic"
#
# ## Go template used to create the tag value. In order to ease TOML
# ## escaping requirements, you may wish to use single quotes around the
# ## template string.
# template = '{{ .Tag "hostname" }}.{{ .Tag "level" }}'
# # Print all metrics that pass through this filter.
# [[processors.topk]]
# ## How many seconds between aggregations
@ -1636,16 +1741,20 @@
# ## of accumulating the results.
# reset = false
#
# ## Whether bucket values should be accumulated. If set to false, "gt" tag will be added.
# ## Defaults to true.
# cumulative = true
#
# ## Example config that aggregates all fields of the metric.
# # [[aggregators.histogram.config]]
# # ## The set of buckets.
# # ## Right borders of buckets (with +Inf implicitly added).
# # buckets = [0.0, 15.6, 34.5, 49.1, 71.5, 80.5, 94.5, 100.0]
# # ## The name of metric.
# # measurement_name = "cpu"
#
# ## Example config that aggregates only specific fields of the metric.
# # [[aggregators.histogram.config]]
# # ## The set of buckets.
# # ## Right borders of buckets (with +Inf implicitly added).
# # buckets = [0.0, 10.0, 20.0, 30.0, 40.0, 50.0, 60.0, 70.0, 80.0, 90.0, 100.0]
# # ## The name of metric.
# # measurement_name = "diskio"
@ -2123,7 +2232,7 @@
# # Gather health check statuses from services registered in Consul
# [[inputs.consul]]
# ## Consul server address
# # address = "localhost"
# # address = "localhost:8500"
#
# ## URI scheme for the Consul server, one of "http", "https"
# # scheme = "http"
@ -2823,6 +2932,11 @@
# # insecure_skip_verify = true
# # Gets counters from all InfiniBand cards and ports installed
# [[inputs.infiniband]]
# # no configuration
# # Read InfluxDB-formatted JSON metrics from one or more HTTP endpoints
# [[inputs.influxdb]]
# ## Works with InfluxDB debug endpoints out of the box,
@ -3136,7 +3250,7 @@
# # Read status information from one or more Kibana servers
# [[inputs.kibana]]
# ## specify a list of one or more Kibana servers
# ## Specify a list of one or more Kibana servers
# servers = ["http://localhost:5601"]
#
# ## Timeout for HTTP requests
@ -3388,6 +3502,80 @@
# # tagdrop = ["server"]
# # Retrieve data from MODBUS slave devices
# [[inputs.modbus]]
# ## Connection Configuration
# ##
# ## The plugin supports connections to PLCs via MODBUS/TCP or
# ## via serial line communication in binary (RTU) or readable (ASCII) encoding
# ##
# ## Device name
# name = "Device"
#
# ## Slave ID - addresses a MODBUS device on the bus
# ## Range: 0 - 255 [0 = broadcast; 248 - 255 = reserved]
# slave_id = 1
#
# ## Timeout for each request
# timeout = "1s"
#
# # TCP - connect via Modbus/TCP
# controller = "tcp://localhost:502"
#
# # Serial (RS485; RS232)
# #controller = "file:///dev/ttyUSB0"
# #baud_rate = 9600
# #data_bits = 8
# #parity = "N"
# #stop_bits = 1
# #transmission_mode = "RTU"
#
#
# ## Measurements
# ##
#
# ## Digital Variables, Discrete Inputs and Coils
# ## name - the variable name
# ## address - variable address
#
# discrete_inputs = [
# { name = "start", address = [0]},
# { name = "stop", address = [1]},
# { name = "reset", address = [2]},
# { name = "emergency_stop", address = [3]},
# ]
# coils = [
# { name = "motor1_run", address = [0]},
# { name = "motor1_jog", address = [1]},
# { name = "motor1_stop", address = [2]},
# ]
#
# ## Analog Variables, Input Registers and Holding Registers
# ## name - the variable name
# ## byte_order - the ordering of bytes
# ## |---AB, ABCD - Big Endian
# ## |---BA, DCBA - Little Endian
# ## |---BADC - Mid-Big Endian
# ## |---CDAB - Mid-Little Endian
# ## data_type - UINT16, INT16, INT32, UINT32, FLOAT32, FLOAT32-IEEE (the IEEE 754 binary representation)
# ## scale - the final numeric variable representation
# ## address - variable address
#
# holding_registers = [
# { name = "power_factor", byte_order = "AB", data_type = "FLOAT32", scale=0.01, address = [8]},
# { name = "voltage", byte_order = "AB", data_type = "FLOAT32", scale=0.1, address = [0]},
# { name = "energy", byte_order = "ABCD", data_type = "FLOAT32", scale=0.001, address = [5,6]},
# { name = "current", byte_order = "ABCD", data_type = "FLOAT32", scale=0.001, address = [1,2]},
# { name = "frequency", byte_order = "AB", data_type = "FLOAT32", scale=0.1, address = [7]},
# { name = "power", byte_order = "ABCD", data_type = "FLOAT32", scale=0.1, address = [3,4]},
# ]
# input_registers = [
# { name = "tank_level", byte_order = "AB", data_type = "INT16", scale=1.0, address = [0]},
# { name = "tank_ph", byte_order = "AB", data_type = "INT16", scale=1.0, address = [1]},
# { name = "pump1_speed", byte_order = "ABCD", data_type = "INT32", scale=1.0, address = [3,4]},
# ]
# # Read metrics from one or many MongoDB servers
# [[inputs.mongodb]]
# ## An array of URLs of the form:
@ -3415,6 +3603,26 @@
# # insecure_skip_verify = false
# # Read metrics and status information about processes managed by Monit
# [[inputs.monit]]
# ## Monit HTTPD address
# address = "http://127.0.0.1:2812"
#
# ## Username and Password for Monit
# # username = ""
# # password = ""
#
# ## Amount of time allowed to complete the HTTP request
# # timeout = "5s"
#
# ## Optional TLS Config
# # tls_ca = "/etc/telegraf/ca.pem"
# # tls_cert = "/etc/telegraf/cert.pem"
# # tls_key = "/etc/telegraf/key.pem"
# ## Use TLS but skip chain & host verification
# # insecure_skip_verify = false
# # Aggregates the contents of multiple files into a single point
# [[inputs.multifile]]
# ## Base directory where telegraf will look for files.
@ -3772,7 +3980,7 @@
# # timeout = "5ms"
# # A plugin to collect stats from Opensmtpd - a validating, recursive, and caching DNS resolver
# # A plugin to collect stats from Opensmtpd - a validating, recursive, and caching DNS resolver
# [[inputs.opensmtpd]]
# ## If running as a restricted user you can prepend sudo for additional access:
# #use_sudo = false
@ -4672,6 +4880,10 @@
# ## Timeout for SSL connection
# # timeout = "5s"
#
# ## Pass a different name into the TLS request (Server Name Indication)
# ## example: server_name = "myhost.example.org"
# # server_name = ""
#
# ## Optional TLS Config
# # tls_ca = "/etc/telegraf/ca.pem"
# # tls_cert = "/etc/telegraf/cert.pem"
@ -4905,6 +5117,73 @@
# ifstats = "ietf-interfaces:interfaces-state/interface/statistics"
# # Read metrics from one or many ClickHouse servers
# [[inputs.clickhouse]]
# ## Username for authorization on ClickHouse server
# ## example: user = "default""
# username = "default"
#
# ## Password for authorization on ClickHouse server
# ## example: password = "super_secret"
#
# ## HTTP(s) timeout while getting metrics values
# ## The timeout includes connection time, any redirects, and reading the response body.
# ## example: timeout = 1s
# # timeout = 5s
#
# ## List of servers for metrics scraping
# ## metrics scrape via HTTP(s) clickhouse interface
# ## https://clickhouse.tech/docs/en/interfaces/http/
# ## example: servers = ["http://127.0.0.1:8123","https://custom-server.mdb.yandexcloud.net"]
# servers = ["http://127.0.0.1:8123"]
#
# ## If "auto_discovery"" is "true" plugin tries to connect to all servers available in the cluster
# ## with using same "user:password" described in "user" and "password" parameters
# ## and get this server hostname list from "system.clusters" table
# ## see
# ## - https://clickhouse.tech/docs/en/operations/system_tables/#system-clusters
# ## - https://clickhouse.tech/docs/en/operations/server_settings/settings/#server_settings_remote_servers
# ## - https://clickhouse.tech/docs/en/operations/table_engines/distributed/
# ## - https://clickhouse.tech/docs/en/operations/table_engines/replication/#creating-replicated-tables
# ## example: auto_discovery = false
# # auto_discovery = true
#
# ## Filter cluster names in "system.clusters" when "auto_discovery" is "true"
# ## when this filter present then "WHERE cluster IN (...)" filter will apply
# ## please use only full cluster names here, regexp and glob filters is not allowed
# ## for "/etc/clickhouse-server/config.d/remote.xml"
# ## <yandex>
# ## <remote_servers>
# ## <my-own-cluster>
# ## <shard>
# ## <replica><host>clickhouse-ru-1.local</host><port>9000</port></replica>
# ## <replica><host>clickhouse-ru-2.local</host><port>9000</port></replica>
# ## </shard>
# ## <shard>
# ## <replica><host>clickhouse-eu-1.local</host><port>9000</port></replica>
# ## <replica><host>clickhouse-eu-2.local</host><port>9000</port></replica>
# ## </shard>
# ## </my-onw-cluster>
# ## </remote_servers>
# ##
# ## </yandex>
# ##
# ## example: cluster_include = ["my-own-cluster"]
# # cluster_include = []
#
# ## Filter cluster names in "system.clusters" when "auto_discovery" is "true"
# ## when this filter present then "WHERE cluster NOT IN (...)" filter will apply
# ## example: cluster_exclude = ["my-internal-not-discovered-cluster"]
# # cluster_exclude = []
#
# ## Optional TLS Config
# # tls_ca = "/etc/telegraf/ca.pem"
# # tls_cert = "/etc/telegraf/cert.pem"
# # tls_key = "/etc/telegraf/key.pem"
# ## Use TLS but skip chain & host verification
# # insecure_skip_verify = false
# # Read metrics from Google PubSub
# [[inputs.cloud_pubsub]]
# ## Required. Name of Google Cloud Platform (GCP) Project that owns
@ -5071,9 +5350,113 @@
# # insecure_skip_verify = false
# # Influx HTTP write listener
# # Azure Event Hubs service input plugin
# [[inputs.eventhub_consumer]]
# ## The default behavior is to create a new Event Hub client from environment variables.
# ## This requires one of the following sets of environment variables to be set:
# ##
# ## 1) Expected Environment Variables:
# ## - "EVENTHUB_NAMESPACE"
# ## - "EVENTHUB_NAME"
# ## - "EVENTHUB_CONNECTION_STRING"
# ##
# ## 2) Expected Environment Variables:
# ## - "EVENTHUB_NAMESPACE"
# ## - "EVENTHUB_NAME"
# ## - "EVENTHUB_KEY_NAME"
# ## - "EVENTHUB_KEY_VALUE"
#
# ## Uncommenting the option below will create an Event Hub client based solely on the connection string.
# ## This can either be the associated environment variable or hard coded directly.
# # connection_string = ""
#
# ## Set persistence directory to a valid folder to use a file persister instead of an in-memory persister
# # persistence_dir = ""
#
# ## Change the default consumer group
# # consumer_group = ""
#
# ## By default the event hub receives all messages present on the broker, alternative modes can be set below.
# ## The timestamp should be in https://github.com/toml-lang/toml#offset-date-time format (RFC 3339).
# ## The 3 options below only apply if no valid persister is read from memory or file (e.g. first run).
# # from_timestamp =
# # latest = true
#
# ## Set a custom prefetch count for the receiver(s)
# # prefetch_count = 1000
#
# ## Add an epoch to the receiver(s)
# # epoch = 0
#
# ## Change to set a custom user agent, "telegraf" is used by default
# # user_agent = "telegraf"
#
# ## To consume from a specific partition, set the partition_ids option.
# ## An empty array will result in receiving from all partitions.
# # partition_ids = ["0","1"]
#
# ## Max undelivered messages
# # max_undelivered_messages = 1000
#
# ## Set either option below to true to use a system property as timestamp.
# ## You have the choice between EnqueuedTime and IoTHubEnqueuedTime.
# ## It is recommended to use this setting when the data itself has no timestamp.
# # enqueued_time_as_ts = true
# # iot_hub_enqueued_time_as_ts = true
#
# ## Tags or fields to create from keys present in the application property bag.
# ## These could for example be set by message enrichments in Azure IoT Hub.
# # application_property_tags = []
# # application_property_fields = []
#
# ## Tag or field name to use for metadata
# ## By default all metadata is disabled
# # sequence_number_field = "SequenceNumber"
# # enqueued_time_field = "EnqueuedTime"
# # offset_field = "Offset"
# # partition_id_tag = "PartitionID"
# # partition_key_tag = "PartitionKey"
# # iot_hub_device_connection_id_tag = "IoTHubDeviceConnectionID"
# # iot_hub_auth_generation_id_tag = "IoTHubAuthGenerationID"
# # iot_hub_connection_auth_method_tag = "IoTHubConnectionAuthMethod"
# # iot_hub_connection_module_id_tag = "IoTHubConnectionModuleID"
# # iot_hub_enqueued_time_field = "IoTHubEnqueuedTime"
#
# ## Data format to consume.
# ## Each data format has its own unique set of configuration options, read
# ## more about them here:
# ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
# data_format = "influx"
# # Run executable as long-running input plugin
# [[inputs.execd]]
# ## Program to run as daemon
# command = ["telegraf-smartctl", "-d", "/dev/sda"]
#
# ## Define how the process is signaled on each collection interval.
# ## Valid values are:
# ## "none" : Do not signal anything.
# ## The process must output metrics by itself.
# ## "STDIN" : Send a newline on STDIN.
# ## "SIGHUP" : Send a HUP signal. Not available on Windows.
# ## "SIGUSR1" : Send a USR1 signal. Not available on Windows.
# ## "SIGUSR2" : Send a USR2 signal. Not available on Windows.
# signal = "none"
#
# ## Delay before the process is restarted after an unexpected termination
# restart_delay = "10s"
#
# ## Data format to consume.
# ## Each data format has its own unique set of configuration options, read
# ## more about them here:
# ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
# data_format = "influx"
# # Accept metrics over InfluxDB 1.x HTTP API
# [[inputs.http_listener]]
# ## Address and port to host HTTP listener on
# ## Address and port to host InfluxDB listener on
# service_address = ":8186"
#
# ## maximum duration before timing out read of the request
@ -5081,14 +5464,9 @@
# ## maximum duration before timing out write of the response
# write_timeout = "10s"
#
# ## Maximum allowed http request body size in bytes.
# ## 0 means to use the default of 524,288,000 bytes (500 mebibytes)
# max_body_size = "500MiB"
#
# ## Maximum line size allowed to be sent in bytes.
# ## 0 means to use the default of 65536 bytes (64 kibibytes)
# max_line_size = "64KiB"
#
# ## Maximum allowed HTTP request body size in bytes.
# ## 0 means to use the default of 32MiB.
# max_body_size = "32MiB"
#
# ## Optional tag name used to store the database.
# ## If the write has a database in the query string then it will be kept in this tag name.
@ -5154,9 +5532,9 @@
# data_format = "influx"
# # Influx HTTP write listener
# # Accept metrics over InfluxDB 1.x HTTP API
# [[inputs.influxdb_listener]]
# ## Address and port to host HTTP listener on
# ## Address and port to host InfluxDB listener on
# service_address = ":8186"
#
# ## maximum duration before timing out read of the request
@ -5164,14 +5542,9 @@
# ## maximum duration before timing out write of the response
# write_timeout = "10s"
#
# ## Maximum allowed http request body size in bytes.
# ## 0 means to use the default of 524,288,000 bytes (500 mebibytes)
# max_body_size = "500MiB"
#
# ## Maximum line size allowed to be sent in bytes.
# ## 0 means to use the default of 65536 bytes (64 kibibytes)
# max_line_size = "64KiB"
#
# ## Maximum allowed HTTP request body size in bytes.
# ## 0 means to use the default of 32MiB.
# max_body_size = "32MiB"
#
# ## Optional tag name used to store the database.
# ## If the write has a database in the query string then it will be kept in this tag name.
@ -5393,6 +5766,14 @@
# table_name = "default"
# # Read metrics off Arista LANZ, via socket
# [[inputs.lanz]]
# ## URL to Arista LANZ endpoint
# servers = [
# "tcp://127.0.0.1:50001"
# ]
# # Stream and parse log file(s).
# [[inputs.logparser]]
# ## Log files to parse.
@ -5529,6 +5910,9 @@
# # username = ""
# # password = ""
#
# ## Optional NATS 2.0 and NATS NGS compatible user credentials
# # credentials = "/etc/telegraf/nats.creds"
#
# ## Use Transport Layer Security
# # secure = false
#
@ -5595,10 +5979,10 @@
# # Read metrics from one or many pgbouncer servers
# [[inputs.pgbouncer]]
# ## specify address via a url matching:
# ## postgres://[pqgotest[:password]]@host:port[/dbname]\
# ## postgres://[pqgotest[:password]]@localhost[/dbname]\
# ## ?sslmode=[disable|verify-ca|verify-full]
# ## or a simple string:
# ## host=localhost port=5432 user=pqotest password=... sslmode=... dbname=app_production
# ## host=localhost user=pqotest password=... sslmode=... dbname=app_production
# ##
# ## All connection parameters are optional.
# ##
@ -5608,10 +5992,10 @@
# # Read metrics from one or many postgresql servers
# [[inputs.postgresql]]
# ## specify address via a url matching:
# ## postgres://[pqgotest[:password]]@host:port[/dbname]\
# ## postgres://[pqgotest[:password]]@localhost[/dbname]\
# ## ?sslmode=[disable|verify-ca|verify-full]
# ## or a simple string:
# ## host=localhost port=5432 user=pqotest password=... sslmode=... dbname=app_production
# ## host=localhost user=pqotest password=... sslmode=... dbname=app_production
# ##
# ## All connection parameters are optional.
# ##
@ -5620,7 +6004,7 @@
# ## connection with the server and doesn't restrict the databases we are trying
# ## to grab metrics for.
# ##
# address = "host=localhost port=5432 user=postgres sslmode=disable"
# address = "host=localhost user=postgres sslmode=disable"
# ## A custom name for the database that will be used as the "server" tag in the
# ## measurement output. If not specified, a default one generated from
# ## the connection address is used.
@ -5643,10 +6027,10 @@
# # Read metrics from one or many postgresql servers
# [[inputs.postgresql_extensible]]
# ## specify address via a url matching:
# ## postgres://[pqgotest[:password]]@host:port[/dbname]\
# ## postgres://[pqgotest[:password]]@localhost[/dbname]\
# ## ?sslmode=[disable|verify-ca|verify-full]
# ## or a simple string:
# ## host=localhost port=5432 user=pqotest password=... sslmode=... dbname=app_production
# ## host=localhost user=pqotest password=... sslmode=... dbname=app_production
# #
# ## All connection parameters are optional. #
# ## Without the dbname parameter, the driver will default to a database
@ -5654,7 +6038,7 @@
# ## connection with the server and doesn't restrict the databases we are trying
# ## to grab metrics for.
# #
# address = "host=localhost port=5432 user=postgres sslmode=disable"
# address = "host=localhost user=postgres sslmode=disable"
#
# ## connection configuration.
# ## maxlifetime - specify the maximum lifetime of a connection.
@ -5740,6 +6124,11 @@
# ## Restricts Kubernetes monitoring to a single namespace
# ## ex: monitor_kubernetes_pods_namespace = "default"
# # monitor_kubernetes_pods_namespace = ""
# # label selector to target pods which have the label
# # kubernetes_label_selector = "env=dev,app=nginx"
# # field selector to target pods
# # eg. To scrape pods on a specific node
# # kubernetes_field_selector = "spec.nodeName=$HOSTNAME"
#
# ## Use bearer token for authorization. ('bearer_token' takes priority)
# # bearer_token = "/path/to/bearer/token"
@ -5762,6 +6151,19 @@
# # insecure_skip_verify = false
# # SFlow V5 Protocol Listener
# [[inputs.sflow]]
# ## Address to listen for sFlow packets.
# ## example: service_address = "udp://:6343"
# ## service_address = "udp4://:6343"
# ## service_address = "udp6://:6343"
# service_address = "udp://:6343"
#
# ## Set the size of the operating system's receive buffer.
# ## example: read_buffer_size = "64KiB"
# # read_buffer_size = ""
# # Receive SNMP traps
# [[inputs.snmp_trap]]
# ## Transport, local address, and port to listen on. Transport must
@ -6135,11 +6537,6 @@
# # collect_concurrency = 1
# # discover_concurrency = 1
#
# ## whether or not to force discovery of new objects on initial gather call before collecting metrics
# ## when true for large environments this may cause errors for time elapsed while collecting metrics
# ## when false (default) the first collection cycle may result in no or limited metrics while objects are discovered
# # force_discover_on_init = false
#
# ## the interval before (re)discovering objects subject to metrics collection (default: 300s)
# # object_discovery_interval = "300s"
#