diff --git a/etc/telegraf.conf b/etc/telegraf.conf index 2ad0bcbae..9f95a9ca5 100644 --- a/etc/telegraf.conf +++ b/etc/telegraf.conf @@ -66,7 +66,7 @@ debug = false ## Run telegraf in quiet mode (error log messages only). quiet = false - ## Specify the log file name. The empty string means to log to stdout. + ## Specify the log file name. The empty string means to log to stderr. logfile = "" ## Override default hostname, if empty use os.Hostname() @@ -441,6 +441,30 @@ +############################################################################### +# PROCESSOR PLUGINS # +############################################################################### + +# # Print all metrics that pass through this filter. +# [[processors.printer]] + + + +############################################################################### +# AGGREGATOR PLUGINS # +############################################################################### + +# # Keep the aggregate min/max of each metric passing through. +# [[aggregators.minmax]] +# ## General Aggregator Arguments: +# ## The period on which to flush & clear the aggregator. +# period = "30s" +# ## If true, the original metric will be dropped by the +# ## aggregator and will not get sent to the output plugins. +# drop_original = false + + + ############################################################################### # INPUT PLUGINS # ############################################################################### @@ -582,15 +606,18 @@ # # Read specific statistics per cgroup # [[inputs.cgroup]] -# ## Directories in which to look for files, globs are supported. -# # paths = [ -# # "/cgroup/memory", -# # "/cgroup/memory/child1", -# # "/cgroup/memory/child2/*", -# # ] -# ## cgroup stat fields, as file names, globs are supported. -# ## these file names are appended to each path from above. -# # files = ["memory.*usage*", "memory.limit_in_bytes"] +# ## Directories in which to look for files, globs are supported. +# ## Consider restricting paths to the set of cgroups you really +# ## want to monitor if you have a large number of cgroups, to avoid +# ## any cardinality issues. +# # paths = [ +# # "/cgroup/memory", +# # "/cgroup/memory/child1", +# # "/cgroup/memory/child2/*", +# # ] +# ## cgroup stat fields, as file names, globs are supported. +# ## these file names are appended to each path from above. +# # files = ["memory.*usage*", "memory.limit_in_bytes"] # # Pull Metric Statistics from Amazon CloudWatch @@ -850,12 +877,15 @@ # ## An array of address to gather stats about. Specify an ip on hostname # ## with optional port. ie localhost, 10.10.3.33:1936, etc. # ## Make sure you specify the complete path to the stats endpoint -# ## ie 10.10.3.33:1936/haproxy?stats +# ## including the protocol, ie http://10.10.3.33:1936/haproxy?stats # # # ## If no servers are specified, then default to 127.0.0.1:1936/haproxy?stats # servers = ["http://myhaproxy.com:1936/haproxy?stats"] -# ## Or you can also use local socket -# ## servers = ["socket:/run/haproxy/admin.sock"] +# ## +# ## You can also use local socket with standard wildcard globbing. +# ## Server address not starting with 'http' will be treated as a possible +# ## socket, so both examples below are valid. +# ## servers = ["socket:/run/haproxy/admin.sock", "/run/haproxy/*.sock"] # # HTTP/HTTPS request given an address a method and a timeout @@ -1000,6 +1030,22 @@ # attribute = "LoadedClassCount,UnloadedClassCount,TotalLoadedClassCount" +# # Read metrics from the kubernetes kubelet api +# [[inputs.kubernetes]] +# ## URL for the kubelet +# url = "http://1.1.1.1:10255" +# +# ## Use bearer token for authorization +# # bearer_token = /path/to/bearer/token +# +# ## Optional SSL Config +# # ssl_ca = /path/to/cafile +# # ssl_cert = /path/to/certfile +# # ssl_key = /path/to/keyfile +# ## Use SSL but skip chain & host verification +# # insecure_skip_verify = false + + # # Read metrics from a LeoFS Server via SNMP # [[inputs.leofs]] # ## An array of URI to gather stats about LeoFS. @@ -1119,13 +1165,13 @@ # ## gather metrics from SHOW BINARY LOGS command output # gather_binary_logs = false # # -# ## gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMART_BY_TABLE +# ## gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMARY_BY_TABLE # gather_table_io_waits = false # # # ## gather metrics from PERFORMANCE_SCHEMA.TABLE_LOCK_WAITS # gather_table_lock_waits = false # # -# ## gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMART_BY_INDEX_USAGE +# ## gather metrics from PERFORMANCE_SCHEMA.TABLE_IO_WAITS_SUMMARY_BY_INDEX_USAGE # gather_index_io_waits = false # # # ## gather metrics from PERFORMANCE_SCHEMA.EVENT_WAITS @@ -1247,13 +1293,13 @@ # ## urls to ping # urls = ["www.google.com"] # required # ## number of pings to send per collection (ping -c ) -# count = 1 # required +# # count = 1 # ## interval, in s, at which to ping. 0 == default (ping -i ) -# ping_interval = 0.0 +# # ping_interval = 1.0 # ## per-ping timeout, in s. 0 == no timeout (ping -W ) -# timeout = 1.0 +# # timeout = 1.0 # ## interface to send ping from (ping -I ) -# interface = "" +# # interface = "" # # Read metrics from one or many postgresql servers @@ -1681,9 +1727,18 @@ # ## Address and port to host HTTP listener on # service_address = ":8186" # -# ## timeouts +# ## maximum duration before timing out read of the request # read_timeout = "10s" +# ## maximum duration before timing out write of the response # write_timeout = "10s" +# +# ## Maximum allowed http request body size in bytes. +# ## 0 means to use the default of 536,870,912 bytes (500 mebibytes) +# max_body_size = 0 +# +# ## Maximum line size allowed to be sent in bytes. +# ## 0 means to use the default of 65536 bytes (64 kibibytes) +# max_line_size = 0 # # Read metrics from Kafka topic(s) @@ -1778,13 +1833,18 @@ # # Read metrics from NATS subject(s) # [[inputs.nats_consumer]] # ## urls of NATS servers -# servers = ["nats://localhost:4222"] +# # servers = ["nats://localhost:4222"] # ## Use Transport Layer Security -# secure = false +# # secure = false # ## subject(s) to consume -# subjects = ["telegraf"] +# # subjects = ["telegraf"] # ## name a queue group -# queue_group = "telegraf_consumers" +# # queue_group = "telegraf_consumers" +# +# ## Sets the limits for pending msgs and bytes for each subscription +# ## These shouldn't need to be adjusted except in very high throughput scenarios +# # pending_message_limit = 65536 +# # pending_bytes_limit = 67108864 # # ## Data format to consume. # ## Each data format has it's own unique set of configuration options, read @@ -1871,14 +1931,14 @@ # # Generic TCP listener # [[inputs.tcp_listener]] # ## Address and port to host TCP listener on -# service_address = ":8094" +# # service_address = ":8094" # # ## Number of TCP messages allowed to queue up. Once filled, the # ## TCP listener will start dropping packets. -# allowed_pending_messages = 10000 +# # allowed_pending_messages = 10000 # # ## Maximum number of concurrent TCP connections to allow -# max_tcp_connections = 250 +# # max_tcp_connections = 250 # # ## Data format to consume. # ## Each data format has it's own unique set of configuration options, read @@ -1890,11 +1950,11 @@ # # Generic UDP listener # [[inputs.udp_listener]] # ## Address and port to host UDP listener on -# service_address = ":8092" +# # service_address = ":8092" # # ## Number of UDP messages allowed to queue up. Once filled, the # ## UDP listener will start dropping packets. -# allowed_pending_messages = 10000 +# # allowed_pending_messages = 10000 # # ## Data format to consume. # ## Each data format has it's own unique set of configuration options, read