Seems to be a toml parse bug around triple pounds

This commit is contained in:
Cameron Sparr 2016-02-18 14:26:51 -07:00
parent 7def6663bd
commit 8d2e5f0bda
75 changed files with 502 additions and 502 deletions

View File

@ -114,10 +114,10 @@ creating the `Parser` object.
You should also add the following to your SampleConfig() return: You should also add the following to your SampleConfig() return:
```toml ```toml
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
``` ```
@ -244,10 +244,10 @@ instantiating and creating the `Serializer` object.
You should also add the following to your SampleConfig() return: You should also add the following to your SampleConfig() return:
```toml ```toml
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
``` ```

View File

@ -25,19 +25,19 @@ example, in the exec plugin:
```toml ```toml
[[inputs.exec]] [[inputs.exec]]
### Commands array ## Commands array
commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"] commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"]
### measurement name suffix (for separating different commands) ## measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "json" data_format = "json"
### Additional configuration options go here ## Additional configuration options go here
``` ```
Each data_format has an additional set of configuration options available, which Each data_format has an additional set of configuration options available, which
@ -52,16 +52,16 @@ metrics are parsed directly into Telegraf metrics.
```toml ```toml
[[inputs.exec]] [[inputs.exec]]
### Commands array ## Commands array
commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"] commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"]
### measurement name suffix (for separating different commands) ## measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
``` ```
@ -97,19 +97,19 @@ For example, if you had this configuration:
```toml ```toml
[[inputs.exec]] [[inputs.exec]]
### Commands array ## Commands array
commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"] commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"]
### measurement name suffix (for separating different commands) ## measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "json" data_format = "json"
### List of tag names to extract from top-level of JSON server response ## List of tag names to extract from top-level of JSON server response
tag_keys = [ tag_keys = [
"my_tag_1", "my_tag_1",
"my_tag_2" "my_tag_2"
@ -241,30 +241,30 @@ There are many more options available,
```toml ```toml
[[inputs.exec]] [[inputs.exec]]
### Commands array ## Commands array
commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"] commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"]
### measurement name suffix (for separating different commands) ## measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Data format to consume. This can be "json", "influx" or "graphite" (line-protocol) ## Data format to consume. This can be "json", "influx" or "graphite" (line-protocol)
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "graphite" data_format = "graphite"
### This string will be used to join the matched values. ## This string will be used to join the matched values.
separator = "_" separator = "_"
### Each template line requires a template pattern. It can have an optional ## Each template line requires a template pattern. It can have an optional
### filter before the template and separated by spaces. It can also have optional extra ## filter before the template and separated by spaces. It can also have optional extra
### tags following the template. Multiple tags should be separated by commas and no spaces ## tags following the template. Multiple tags should be separated by commas and no spaces
### similar to the line protocol format. There can be only one default template. ## similar to the line protocol format. There can be only one default template.
### Templates support below format: ## Templates support below format:
### 1. filter + template ## 1. filter + template
### 2. filter + template + extra tag ## 2. filter + template + extra tag
### 3. filter + template with field key ## 3. filter + template with field key
### 4. default template ## 4. default template
templates = [ templates = [
"*.app env.service.resource.measurement", "*.app env.service.resource.measurement",
"stats.* .host.measurement* region=us-west,agent=sensu", "stats.* .host.measurement* region=us-west,agent=sensu",

View File

@ -26,16 +26,16 @@ config option, for example, in the `file` output plugin:
```toml ```toml
[[outputs.file]] [[outputs.file]]
### Files to write to, "stdout" is a specially handled file. ## Files to write to, "stdout" is a specially handled file.
files = ["stdout"] files = ["stdout"]
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
### Additional configuration options go here ## Additional configuration options go here
``` ```
Each data_format has an additional set of configuration options available, which Each data_format has an additional set of configuration options available, which
@ -50,13 +50,13 @@ metrics are serialized directly into InfluxDB line-protocol.
```toml ```toml
[[outputs.file]] [[outputs.file]]
### Files to write to, "stdout" is a specially handled file. ## Files to write to, "stdout" is a specially handled file.
files = ["stdout", "/tmp/metrics.out"] files = ["stdout", "/tmp/metrics.out"]
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
``` ```
@ -84,13 +84,13 @@ tars.cpu-total.us-east-1.cpu.usage_idle 98.09 1455320690
```toml ```toml
[[outputs.file]] [[outputs.file]]
### Files to write to, "stdout" is a specially handled file. ## Files to write to, "stdout" is a specially handled file.
files = ["stdout", "/tmp/metrics.out"] files = ["stdout", "/tmp/metrics.out"]
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
prefix = "telegraf" prefix = "telegraf"

View File

@ -16,37 +16,37 @@
# Configuration for telegraf agent # Configuration for telegraf agent
[agent] [agent]
### Default data collection interval for all inputs ## Default data collection interval for all inputs
interval = "10s" interval = "10s"
### Rounds collection interval to 'interval' ## Rounds collection interval to 'interval'
### ie, if interval="10s" then always collect on :00, :10, :20, etc. ## ie, if interval="10s" then always collect on :00, :10, :20, etc.
round_interval = true round_interval = true
### Telegraf will cache metric_buffer_limit metrics for each output, and will ## Telegraf will cache metric_buffer_limit metrics for each output, and will
### flush this buffer on a successful write. ## flush this buffer on a successful write.
metric_buffer_limit = 10000 metric_buffer_limit = 10000
### Flush the buffer whenever full, regardless of flush_interval. ## Flush the buffer whenever full, regardless of flush_interval.
flush_buffer_when_full = true flush_buffer_when_full = true
### Collection jitter is used to jitter the collection by a random amount. ## Collection jitter is used to jitter the collection by a random amount.
### Each plugin will sleep for a random time within jitter before collecting. ## Each plugin will sleep for a random time within jitter before collecting.
### This can be used to avoid many plugins querying things like sysfs at the ## This can be used to avoid many plugins querying things like sysfs at the
### same time, which can have a measurable effect on the system. ## same time, which can have a measurable effect on the system.
collection_jitter = "0s" collection_jitter = "0s"
### Default flushing interval for all outputs. You shouldn't set this below ## Default flushing interval for all outputs. You shouldn't set this below
### interval. Maximum flush_interval will be flush_interval + flush_jitter ## interval. Maximum flush_interval will be flush_interval + flush_jitter
flush_interval = "10s" flush_interval = "10s"
### Jitter the flush interval by a random amount. This is primarily to avoid ## Jitter the flush interval by a random amount. This is primarily to avoid
### large write spikes for users running a large number of telegraf instances. ## large write spikes for users running a large number of telegraf instances.
### ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s
flush_jitter = "0s" flush_jitter = "0s"
### Run telegraf in debug mode ## Run telegraf in debug mode
debug = false debug = false
### Run telegraf in quiet mode ## Run telegraf in quiet mode
quiet = false quiet = false
### Override default hostname, if empty use os.Hostname() ## Override default hostname, if empty use os.Hostname()
hostname = "" hostname = ""

View File

@ -153,37 +153,37 @@ var header = `##################################################################
# Configuration for telegraf agent # Configuration for telegraf agent
[agent] [agent]
### Default data collection interval for all inputs ## Default data collection interval for all inputs
interval = "10s" interval = "10s"
### Rounds collection interval to 'interval' ## Rounds collection interval to 'interval'
### ie, if interval="10s" then always collect on :00, :10, :20, etc. ## ie, if interval="10s" then always collect on :00, :10, :20, etc.
round_interval = true round_interval = true
### Telegraf will cache metric_buffer_limit metrics for each output, and will ## Telegraf will cache metric_buffer_limit metrics for each output, and will
### flush this buffer on a successful write. ## flush this buffer on a successful write.
metric_buffer_limit = 10000 metric_buffer_limit = 10000
### Flush the buffer whenever full, regardless of flush_interval. ## Flush the buffer whenever full, regardless of flush_interval.
flush_buffer_when_full = true flush_buffer_when_full = true
### Collection jitter is used to jitter the collection by a random amount. ## Collection jitter is used to jitter the collection by a random amount.
### Each plugin will sleep for a random time within jitter before collecting. ## Each plugin will sleep for a random time within jitter before collecting.
### This can be used to avoid many plugins querying things like sysfs at the ## This can be used to avoid many plugins querying things like sysfs at the
### same time, which can have a measurable effect on the system. ## same time, which can have a measurable effect on the system.
collection_jitter = "0s" collection_jitter = "0s"
### Default flushing interval for all outputs. You shouldn't set this below ## Default flushing interval for all outputs. You shouldn't set this below
### interval. Maximum flush_interval will be flush_interval + flush_jitter ## interval. Maximum flush_interval will be flush_interval + flush_jitter
flush_interval = "10s" flush_interval = "10s"
### Jitter the flush interval by a random amount. This is primarily to avoid ## Jitter the flush interval by a random amount. This is primarily to avoid
### large write spikes for users running a large number of telegraf instances. ## large write spikes for users running a large number of telegraf instances.
### ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s ## ie, a jitter of 5s and interval 10s means flushes will happen every 10-15s
flush_jitter = "0s" flush_jitter = "0s"
### Run telegraf in debug mode ## Run telegraf in debug mode
debug = false debug = false
### Run telegraf in quiet mode ## Run telegraf in quiet mode
quiet = false quiet = false
### Override default hostname, if empty use os.Hostname() ## Override default hostname, if empty use os.Hostname()
hostname = "" hostname = ""

View File

@ -104,9 +104,9 @@ type Aerospike struct {
} }
var sampleConfig = ` var sampleConfig = `
### Aerospike servers to connect to (with port) ## Aerospike servers to connect to (with port)
### This plugin will query all namespaces the aerospike ## This plugin will query all namespaces the aerospike
### server has configured and get stats for them. ## server has configured and get stats for them.
servers = ["localhost:3000"] servers = ["localhost:3000"]
` `

View File

@ -20,7 +20,7 @@ type Apache struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of Apache status URI to gather stats. ## An array of Apache status URI to gather stats.
urls = ["http://localhost/server-status?auto"] urls = ["http://localhost/server-status?auto"]
` `

View File

@ -18,13 +18,13 @@ type Bcache struct {
} }
var sampleConfig = ` var sampleConfig = `
### Bcache sets path ## Bcache sets path
### If not specified, then default is: ## If not specified, then default is:
bcachePath = "/sys/fs/bcache" bcachePath = "/sys/fs/bcache"
### By default, telegraf gather stats for all bcache devices ## By default, telegraf gather stats for all bcache devices
### Setting devices will restrict the stats to the specified ## Setting devices will restrict the stats to the specified
### bcache devices. ## bcache devices.
bcacheDevs = ["bcache0"] bcacheDevs = ["bcache0"]
` `

View File

@ -75,8 +75,8 @@ func (*CouchDB) Description() string {
func (*CouchDB) SampleConfig() string { func (*CouchDB) SampleConfig() string {
return ` return `
### Works with CouchDB stats endpoints out of the box ## Works with CouchDB stats endpoints out of the box
### Multiple HOSTs from which to read CouchDB stats: ## Multiple HOSTs from which to read CouchDB stats:
hosts = ["http://localhost:8086/_stats"] hosts = ["http://localhost:8086/_stats"]
` `
} }

View File

@ -22,11 +22,11 @@ type Disque struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of URI to gather stats about. Specify an ip or hostname ## An array of URI to gather stats about. Specify an ip or hostname
### with optional port and password. ie disque://localhost, disque://10.10.3.33:18832, ## with optional port and password. ie disque://localhost, disque://10.10.3.33:18832,
### 10.0.0.1:10000, etc. ## 10.0.0.1:10000, etc.
### If no servers are specified, then localhost is used as the host. ## If no servers are specified, then localhost is used as the host.
servers = ["localhost"] servers = ["localhost"]
` `

View File

@ -21,11 +21,11 @@ type Docker struct {
} }
var sampleConfig = ` var sampleConfig = `
### Docker Endpoint ## Docker Endpoint
### To use TCP, set endpoint = "tcp://[ip]:[port]" ## To use TCP, set endpoint = "tcp://[ip]:[port]"
### To use environment variables (ie, docker-machine), set endpoint = "ENV" ## To use environment variables (ie, docker-machine), set endpoint = "ENV"
endpoint = "unix:///var/run/docker.sock" endpoint = "unix:///var/run/docker.sock"
### Only collect metrics for these containers, collect all if empty ## Only collect metrics for these containers, collect all if empty
container_names = [] container_names = []
` `

View File

@ -24,13 +24,13 @@ func (d *Dovecot) Description() string {
} }
var sampleConfig = ` var sampleConfig = `
### specify dovecot servers via an address:port list ## specify dovecot servers via an address:port list
### e.g. ## e.g.
### localhost:24242 ## localhost:24242
### ##
### If no servers are specified, then localhost is used as the host. ## If no servers are specified, then localhost is used as the host.
servers = ["localhost:24242"] servers = ["localhost:24242"]
### Only collect metrics for these domains, collect all if empty ## Only collect metrics for these domains, collect all if empty
domains = [] domains = []
` `

View File

@ -59,14 +59,14 @@ type indexHealth struct {
} }
const sampleConfig = ` const sampleConfig = `
### specify a list of one or more Elasticsearch servers ## specify a list of one or more Elasticsearch servers
servers = ["http://localhost:9200"] servers = ["http://localhost:9200"]
### set local to false when you want to read the indices stats from all nodes ## set local to false when you want to read the indices stats from all nodes
### within the cluster ## within the cluster
local = true local = true
### set cluster_health to true when you want to also obtain cluster level stats ## set cluster_health to true when you want to also obtain cluster level stats
cluster_health = false cluster_health = false
` `

View File

@ -37,19 +37,19 @@ and strings will be ignored.
# measurement name suffix (for separating different commands) # measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Below configuration will be used for data_format = "graphite", can be ignored for other data_format ## Below configuration will be used for data_format = "graphite", can be ignored for other data_format
### If matching multiple measurement files, this string will be used to join the matched values. ## If matching multiple measurement files, this string will be used to join the matched values.
#separator = "." #separator = "."
### Each template line requires a template pattern. It can have an optional ## Each template line requires a template pattern. It can have an optional
### filter before the template and separated by spaces. It can also have optional extra ## filter before the template and separated by spaces. It can also have optional extra
### tags following the template. Multiple tags should be separated by commas and no spaces ## tags following the template. Multiple tags should be separated by commas and no spaces
### similar to the line protocol format. The can be only one default template. ## similar to the line protocol format. The can be only one default template.
### Templates support below format: ## Templates support below format:
### 1. filter + template ## 1. filter + template
### 2. filter + template + extra tag ## 2. filter + template + extra tag
### 3. filter + template with field key ## 3. filter + template with field key
### 4. default template ## 4. default template
#templates = [ #templates = [
# "*.app env.service.resource.measurement", # "*.app env.service.resource.measurement",
# "stats.* .host.measurement* region=us-west,agent=sensu", # "stats.* .host.measurement* region=us-west,agent=sensu",
@ -141,19 +141,19 @@ We can also change the data_format to "graphite" to use the metrics collecting s
# measurement name suffix (for separating different commands) # measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Below configuration will be used for data_format = "graphite", can be ignored for other data_format ## Below configuration will be used for data_format = "graphite", can be ignored for other data_format
### If matching multiple measurement files, this string will be used to join the matched values. ## If matching multiple measurement files, this string will be used to join the matched values.
separator = "." separator = "."
### Each template line requires a template pattern. It can have an optional ## Each template line requires a template pattern. It can have an optional
### filter before the template and separated by spaces. It can also have optional extra ## filter before the template and separated by spaces. It can also have optional extra
### tags following the template. Multiple tags should be separated by commas and no spaces ## tags following the template. Multiple tags should be separated by commas and no spaces
### similar to the line protocol format. The can be only one default template. ## similar to the line protocol format. The can be only one default template.
### Templates support below format: ## Templates support below format:
### 1. filter + template ## 1. filter + template
### 2. filter + template + extra tag ## 2. filter + template + extra tag
### 3. filter + template with field key ## 3. filter + template with field key
### 4. default template ## 4. default template
templates = [ templates = [
"*.app env.service.resource.measurement", "*.app env.service.resource.measurement",
"stats.* .host.measurement* region=us-west,agent=sensu", "stats.* .host.measurement* region=us-west,agent=sensu",

View File

@ -14,16 +14,16 @@ import (
) )
const sampleConfig = ` const sampleConfig = `
### Commands array ## Commands array
commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"] commands = ["/tmp/test.sh", "/usr/bin/mycollector --foo=bar"]
### measurement name suffix (for separating different commands) ## measurement name suffix (for separating different commands)
name_suffix = "_mycollector" name_suffix = "_mycollector"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -31,7 +31,7 @@ func NewGithubWebhooks() *GithubWebhooks {
func (gh *GithubWebhooks) SampleConfig() string { func (gh *GithubWebhooks) SampleConfig() string {
return ` return `
### Address and port to host Webhook listener on ## Address and port to host Webhook listener on
service_address = ":1618" service_address = ":1618"
` `
} }

View File

@ -86,13 +86,13 @@ type haproxy struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of address to gather stats about. Specify an ip on hostname ## An array of address to gather stats about. Specify an ip on hostname
### with optional port. ie localhost, 10.10.3.33:1936, etc. ## with optional port. ie localhost, 10.10.3.33:1936, etc.
### If no servers are specified, then default to 127.0.0.1:1936 ## If no servers are specified, then default to 127.0.0.1:1936
servers = ["http://myhaproxy.com:1936", "http://anotherhaproxy.com:1936"] servers = ["http://myhaproxy.com:1936", "http://anotherhaproxy.com:1936"]
### Or you can also use local socket(not work yet) ## Or you can also use local socket(not work yet)
### servers = ["socket://run/haproxy/admin.sock"] ## servers = ["socket://run/haproxy/admin.sock"]
` `
func (r *haproxy) SampleConfig() string { func (r *haproxy) SampleConfig() string {

View File

@ -47,33 +47,33 @@ func (c RealHTTPClient) MakeRequest(req *http.Request) (*http.Response, error) {
} }
var sampleConfig = ` var sampleConfig = `
### NOTE This plugin only reads numerical measurements, strings and booleans ## NOTE This plugin only reads numerical measurements, strings and booleans
### will be ignored. ## will be ignored.
### a name for the service being polled ## a name for the service being polled
name = "webserver_stats" name = "webserver_stats"
### URL of each server in the service's cluster ## URL of each server in the service's cluster
servers = [ servers = [
"http://localhost:9999/stats/", "http://localhost:9999/stats/",
"http://localhost:9998/stats/", "http://localhost:9998/stats/",
] ]
### HTTP method to use: GET or POST (case-sensitive) ## HTTP method to use: GET or POST (case-sensitive)
method = "GET" method = "GET"
### List of tag names to extract from top-level of JSON server response ## List of tag names to extract from top-level of JSON server response
# tag_keys = [ # tag_keys = [
# "my_tag_1", # "my_tag_1",
# "my_tag_2" # "my_tag_2"
# ] # ]
### HTTP parameters (all values must be strings) ## HTTP parameters (all values must be strings)
[inputs.httpjson.parameters] [inputs.httpjson.parameters]
event_type = "cpu_spike" event_type = "cpu_spike"
threshold = "0.75" threshold = "0.75"
### HTTP Header parameters (all values must be strings) ## HTTP Header parameters (all values must be strings)
# [inputs.httpjson.headers] # [inputs.httpjson.headers]
# X-Auth-Token = "my-xauth-token" # X-Auth-Token = "my-xauth-token"
# apiVersion = "v1" # apiVersion = "v1"

View File

@ -22,11 +22,11 @@ func (*InfluxDB) Description() string {
func (*InfluxDB) SampleConfig() string { func (*InfluxDB) SampleConfig() string {
return ` return `
### Works with InfluxDB debug endpoints out of the box, ## Works with InfluxDB debug endpoints out of the box,
### but other services can use this format too. ## but other services can use this format too.
### See the influxdb plugin's README for more details. ## See the influxdb plugin's README for more details.
### Multiple URLs from which to read InfluxDB-formatted JSON ## Multiple URLs from which to read InfluxDB-formatted JSON
urls = [ urls = [
"http://localhost:8086/debug/vars" "http://localhost:8086/debug/vars"
] ]

View File

@ -46,10 +46,10 @@ type Jolokia struct {
func (j *Jolokia) SampleConfig() string { func (j *Jolokia) SampleConfig() string {
return ` return `
### This is the context root used to compose the jolokia url ## This is the context root used to compose the jolokia url
context = "/jolokia/read" context = "/jolokia/read"
### List of servers exposing jolokia read service ## List of servers exposing jolokia read service
[[inputs.jolokia.servers]] [[inputs.jolokia.servers]]
name = "stable" name = "stable"
host = "192.168.103.2" host = "192.168.103.2"
@ -57,10 +57,10 @@ func (j *Jolokia) SampleConfig() string {
# username = "myuser" # username = "myuser"
# password = "mypassword" # password = "mypassword"
### List of metrics collected on above servers ## List of metrics collected on above servers
### Each metric consists in a name, a jmx path and either ## Each metric consists in a name, a jmx path and either
### a pass or drop slice attribute. ## a pass or drop slice attribute.
### This collect all heap memory usage metrics. ## This collect all heap memory usage metrics.
[[inputs.jolokia.metrics]] [[inputs.jolokia.metrics]]
name = "heap_memory_usage" name = "heap_memory_usage"
jmx = "/java.lang:type=Memory/HeapMemoryUsage" jmx = "/java.lang:type=Memory/HeapMemoryUsage"

View File

@ -11,21 +11,21 @@ from the same topic in parallel.
```toml ```toml
# Read metrics from Kafka topic(s) # Read metrics from Kafka topic(s)
[[inputs.kafka_consumer]] [[inputs.kafka_consumer]]
### topic(s) to consume ## topic(s) to consume
topics = ["telegraf"] topics = ["telegraf"]
### an array of Zookeeper connection strings ## an array of Zookeeper connection strings
zookeeper_peers = ["localhost:2181"] zookeeper_peers = ["localhost:2181"]
### the name of the consumer group ## the name of the consumer group
consumer_group = "telegraf_metrics_consumers" consumer_group = "telegraf_metrics_consumers"
### Maximum number of metrics to buffer between collection intervals ## Maximum number of metrics to buffer between collection intervals
metric_buffer = 100000 metric_buffer = 100000
### Offset (must be either "oldest" or "newest") ## Offset (must be either "oldest" or "newest")
offset = "oldest" offset = "oldest"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
``` ```

View File

@ -44,19 +44,19 @@ type Kafka struct {
} }
var sampleConfig = ` var sampleConfig = `
### topic(s) to consume ## topic(s) to consume
topics = ["telegraf"] topics = ["telegraf"]
### an array of Zookeeper connection strings ## an array of Zookeeper connection strings
zookeeper_peers = ["localhost:2181"] zookeeper_peers = ["localhost:2181"]
### the name of the consumer group ## the name of the consumer group
consumer_group = "telegraf_metrics_consumers" consumer_group = "telegraf_metrics_consumers"
### Offset (must be either "oldest" or "newest") ## Offset (must be either "oldest" or "newest")
offset = "oldest" offset = "oldest"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -132,8 +132,8 @@ var serverTypeMapping = map[string]ServerType{
} }
var sampleConfig = ` var sampleConfig = `
### An array of URI to gather stats about LeoFS. ## An array of URI to gather stats about LeoFS.
### Specify an ip or hostname with port. ie 127.0.0.1:4020 ## Specify an ip or hostname with port. ie 127.0.0.1:4020
servers = ["127.0.0.1:4021"] servers = ["127.0.0.1:4021"]
` `

View File

@ -29,9 +29,9 @@ type Lustre2 struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of /proc globs to search for Lustre stats ## An array of /proc globs to search for Lustre stats
### If not specified, the default will work on Lustre 2.5.x ## If not specified, the default will work on Lustre 2.5.x
### ##
# ost_procfiles = [ # ost_procfiles = [
# "/proc/fs/lustre/obdfilter/*/stats", # "/proc/fs/lustre/obdfilter/*/stats",
# "/proc/fs/lustre/osd-ldiskfs/*/stats" # "/proc/fs/lustre/osd-ldiskfs/*/stats"

View File

@ -17,13 +17,13 @@ type MailChimp struct {
} }
var sampleConfig = ` var sampleConfig = `
### MailChimp API key ## MailChimp API key
### get from https://admin.mailchimp.com/account/api/ ## get from https://admin.mailchimp.com/account/api/
api_key = "" # required api_key = "" # required
### Reports for campaigns sent more than days_old ago will not be collected. ## Reports for campaigns sent more than days_old ago will not be collected.
### 0 means collect all. ## 0 means collect all.
days_old = 0 days_old = 0
### Campaign ID to get, if empty gets all campaigns, this option overrides days_old ## Campaign ID to get, if empty gets all campaigns, this option overrides days_old
# campaign_id = "" # campaign_id = ""
` `

View File

@ -19,8 +19,8 @@ type Memcached struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of address to gather stats about. Specify an ip on hostname ## An array of address to gather stats about. Specify an ip on hostname
### with optional port. ie localhost, 10.0.0.1:11211, etc. ## with optional port. ie localhost, 10.0.0.1:11211, etc.
servers = ["localhost:11211"] servers = ["localhost:11211"]
# unix_sockets = ["/var/run/memcached.sock"] # unix_sockets = ["/var/run/memcached.sock"]
` `

View File

@ -26,11 +26,11 @@ type Ssl struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of URI to gather stats about. Specify an ip or hostname ## An array of URI to gather stats about. Specify an ip or hostname
### with optional port add password. ie, ## with optional port add password. ie,
### mongodb://user:auth_key@10.10.3.30:27017, ## mongodb://user:auth_key@10.10.3.30:27017,
### mongodb://10.10.3.33:18832, ## mongodb://10.10.3.33:18832,
### 10.0.0.1:10000, etc. ## 10.0.0.1:10000, etc.
servers = ["127.0.0.1:27017"] servers = ["127.0.0.1:27017"]
` `

View File

@ -11,34 +11,34 @@ The plugin expects messages in the
# Read metrics from MQTT topic(s) # Read metrics from MQTT topic(s)
[[inputs.mqtt_consumer]] [[inputs.mqtt_consumer]]
servers = ["localhost:1883"] servers = ["localhost:1883"]
### MQTT QoS, must be 0, 1, or 2 ## MQTT QoS, must be 0, 1, or 2
qos = 0 qos = 0
### Topics to subscribe to ## Topics to subscribe to
topics = [ topics = [
"telegraf/host01/cpu", "telegraf/host01/cpu",
"telegraf/+/mem", "telegraf/+/mem",
"sensors/#", "sensors/#",
] ]
### Maximum number of metrics to buffer between collection intervals ## Maximum number of metrics to buffer between collection intervals
metric_buffer = 100000 metric_buffer = 100000
### username and password to connect MQTT server. ## username and password to connect MQTT server.
# username = "telegraf" # username = "telegraf"
# password = "metricsmetricsmetricsmetrics" # password = "metricsmetricsmetricsmetrics"
### Optional SSL Config ## Optional SSL Config
# ssl_ca = "/etc/telegraf/ca.pem" # ssl_ca = "/etc/telegraf/ca.pem"
# ssl_cert = "/etc/telegraf/cert.pem" # ssl_cert = "/etc/telegraf/cert.pem"
# ssl_key = "/etc/telegraf/key.pem" # ssl_key = "/etc/telegraf/key.pem"
### Use SSL but skip chain & host verification ## Use SSL but skip chain & host verification
# insecure_skip_verify = false # insecure_skip_verify = false
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
``` ```

View File

@ -47,31 +47,31 @@ type MQTTConsumer struct {
var sampleConfig = ` var sampleConfig = `
servers = ["localhost:1883"] servers = ["localhost:1883"]
### MQTT QoS, must be 0, 1, or 2 ## MQTT QoS, must be 0, 1, or 2
qos = 0 qos = 0
### Topics to subscribe to ## Topics to subscribe to
topics = [ topics = [
"telegraf/host01/cpu", "telegraf/host01/cpu",
"telegraf/+/mem", "telegraf/+/mem",
"sensors/#", "sensors/#",
] ]
### username and password to connect MQTT server. ## username and password to connect MQTT server.
# username = "telegraf" # username = "telegraf"
# password = "metricsmetricsmetricsmetrics" # password = "metricsmetricsmetricsmetrics"
### Optional SSL Config ## Optional SSL Config
# ssl_ca = "/etc/telegraf/ca.pem" # ssl_ca = "/etc/telegraf/ca.pem"
# ssl_cert = "/etc/telegraf/cert.pem" # ssl_cert = "/etc/telegraf/cert.pem"
# ssl_key = "/etc/telegraf/key.pem" # ssl_key = "/etc/telegraf/key.pem"
### Use SSL but skip chain & host verification ## Use SSL but skip chain & host verification
# insecure_skip_verify = false # insecure_skip_verify = false
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -15,14 +15,14 @@ type Mysql struct {
} }
var sampleConfig = ` var sampleConfig = `
### specify servers via a url matching: ## specify servers via a url matching:
### [username[:password]@][protocol[(address)]]/[?tls=[true|false|skip-verify]] ## [username[:password]@][protocol[(address)]]/[?tls=[true|false|skip-verify]]
### see https://github.com/go-sql-driver/mysql#dsn-data-source-name ## see https://github.com/go-sql-driver/mysql#dsn-data-source-name
### e.g. ## e.g.
### root:passwd@tcp(127.0.0.1:3306)/?tls=false ## root:passwd@tcp(127.0.0.1:3306)/?tls=false
### root@tcp(127.0.0.1:3306)/?tls=false ## root@tcp(127.0.0.1:3306)/?tls=false
### ##
### If no servers are specified, then localhost is used as the host. ## If no servers are specified, then localhost is used as the host.
servers = ["tcp(127.0.0.1:3306)/"] servers = ["tcp(127.0.0.1:3306)/"]
` `

View File

@ -12,20 +12,20 @@ from a NATS cluster in parallel.
```toml ```toml
# Read metrics from NATS subject(s) # Read metrics from NATS subject(s)
[[inputs.nats_consumer]] [[inputs.nats_consumer]]
### urls of NATS servers ## urls of NATS servers
servers = ["nats://localhost:4222"] servers = ["nats://localhost:4222"]
### Use Transport Layer Security ## Use Transport Layer Security
secure = false secure = false
### subject(s) to consume ## subject(s) to consume
subjects = ["telegraf"] subjects = ["telegraf"]
### name a queue group ## name a queue group
queue_group = "telegraf_consumers" queue_group = "telegraf_consumers"
### Maximum number of metrics to buffer between collection intervals ## Maximum number of metrics to buffer between collection intervals
metric_buffer = 100000 metric_buffer = 100000
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
``` ```

View File

@ -46,19 +46,19 @@ type natsConsumer struct {
} }
var sampleConfig = ` var sampleConfig = `
### urls of NATS servers ## urls of NATS servers
servers = ["nats://localhost:4222"] servers = ["nats://localhost:4222"]
### Use Transport Layer Security ## Use Transport Layer Security
secure = false secure = false
### subject(s) to consume ## subject(s) to consume
subjects = ["telegraf"] subjects = ["telegraf"]
### name a queue group ## name a queue group
queue_group = "telegraf_consumers" queue_group = "telegraf_consumers"
### Data format to consume. This can be "json", "influx" or "graphite" ## Data format to consume. This can be "json", "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -27,17 +27,17 @@ func (_ *NetResponse) Description() string {
} }
var sampleConfig = ` var sampleConfig = `
### Protocol, must be "tcp" or "udp" ## Protocol, must be "tcp" or "udp"
protocol = "tcp" protocol = "tcp"
### Server address (default localhost) ## Server address (default localhost)
address = "github.com:80" address = "github.com:80"
### Set timeout (default 1.0 seconds) ## Set timeout (default 1.0 seconds)
timeout = 1.0 timeout = 1.0
### Set read timeout (default 1.0 seconds) ## Set read timeout (default 1.0 seconds)
read_timeout = 1.0 read_timeout = 1.0
### Optional string sent to the server ## Optional string sent to the server
# send = "ssh" # send = "ssh"
### Optional expected string in answer ## Optional expected string in answer
# expect = "ssh" # expect = "ssh"
` `

View File

@ -20,7 +20,7 @@ type Nginx struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of Nginx stub_status URI to gather stats. ## An array of Nginx stub_status URI to gather stats.
urls = ["http://localhost/status"] urls = ["http://localhost/status"]
` `

View File

@ -41,7 +41,7 @@ type NSQ struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of NSQD HTTP API endpoints ## An array of NSQD HTTP API endpoints
endpoints = ["http://localhost:4151"] endpoints = ["http://localhost:4151"]
` `

View File

@ -126,14 +126,14 @@ func (p *process) getUptime() int64 {
} }
var sampleConfig = ` var sampleConfig = `
### Path of passenger-status. ## Path of passenger-status.
### ##
### Plugin gather metric via parsing XML output of passenger-status ## Plugin gather metric via parsing XML output of passenger-status
### More information about the tool: ## More information about the tool:
### https://www.phusionpassenger.com/library/admin/apache/overall_status_report.html ## https://www.phusionpassenger.com/library/admin/apache/overall_status_report.html
### ##
### If no path is specified, then the plugin simply execute passenger-status ## If no path is specified, then the plugin simply execute passenger-status
### hopefully it can be found in your PATH ## hopefully it can be found in your PATH
command = "passenger-status -v --show=xml" command = "passenger-status -v --show=xml"
` `

View File

@ -41,25 +41,25 @@ type phpfpm struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of addresses to gather stats about. Specify an ip or hostname ## An array of addresses to gather stats about. Specify an ip or hostname
### with optional port and path ## with optional port and path
### ##
### Plugin can be configured in three modes (either can be used): ## Plugin can be configured in three modes (either can be used):
### - http: the URL must start with http:// or https://, ie: ## - http: the URL must start with http:// or https://, ie:
### "http://localhost/status" ## "http://localhost/status"
### "http://192.168.130.1/status?full" ## "http://192.168.130.1/status?full"
### ##
### - unixsocket: path to fpm socket, ie: ## - unixsocket: path to fpm socket, ie:
### "/var/run/php5-fpm.sock" ## "/var/run/php5-fpm.sock"
### or using a custom fpm status path: ## or using a custom fpm status path:
### "/var/run/php5-fpm.sock:fpm-custom-status-path" ## "/var/run/php5-fpm.sock:fpm-custom-status-path"
### ##
### - fcgi: the URL must start with fcgi:// or cgi://, and port must be present, ie: ## - fcgi: the URL must start with fcgi:// or cgi://, and port must be present, ie:
### "fcgi://10.0.0.12:9000/status" ## "fcgi://10.0.0.12:9000/status"
### "cgi://10.0.10.12:9001/status" ## "cgi://10.0.10.12:9001/status"
### ##
### Example of multiple gathering from local socket and remove host ## Example of multiple gathering from local socket and remove host
### urls = ["http://192.168.1.20/status", "/tmp/fpm.sock"] ## urls = ["http://192.168.1.20/status", "/tmp/fpm.sock"]
urls = ["http://localhost/status"] urls = ["http://localhost/status"]
` `

View File

@ -44,18 +44,18 @@ func (_ *Ping) Description() string {
} }
var sampleConfig = ` var sampleConfig = `
### NOTE: this plugin forks the ping command. You may need to set capabilities ## NOTE: this plugin forks the ping command. You may need to set capabilities
### via setcap cap_net_raw+p /bin/ping ## via setcap cap_net_raw+p /bin/ping
### urls to ping ## urls to ping
urls = ["www.google.com"] # required urls = ["www.google.com"] # required
### number of pings to send (ping -c <COUNT>) ## number of pings to send (ping -c <COUNT>)
count = 1 # required count = 1 # required
### interval, in s, at which to ping. 0 == default (ping -i <PING_INTERVAL>) ## interval, in s, at which to ping. 0 == default (ping -i <PING_INTERVAL>)
ping_interval = 0.0 ping_interval = 0.0
### ping timeout, in s. 0 == no timeout (ping -t <TIMEOUT>) ## ping timeout, in s. 0 == no timeout (ping -t <TIMEOUT>)
timeout = 0.0 timeout = 0.0
### interface to send ping from (ping -I <INTERFACE>) ## interface to send ping from (ping -I <INTERFACE>)
interface = "" interface = ""
` `

View File

@ -23,22 +23,22 @@ type Postgresql struct {
var ignoredColumns = map[string]bool{"datid": true, "datname": true, "stats_reset": true} var ignoredColumns = map[string]bool{"datid": true, "datname": true, "stats_reset": true}
var sampleConfig = ` var sampleConfig = `
### specify address via a url matching: ## specify address via a url matching:
### postgres://[pqgotest[:password]]@localhost[/dbname]?sslmode=[disable|verify-ca|verify-full] ## postgres://[pqgotest[:password]]@localhost[/dbname]?sslmode=[disable|verify-ca|verify-full]
### or a simple string: ## or a simple string:
### host=localhost user=pqotest password=... sslmode=... dbname=app_production ## host=localhost user=pqotest password=... sslmode=... dbname=app_production
### ##
### All connection parameters are optional. ## All connection parameters are optional.
### ##
### Without the dbname parameter, the driver will default to a database ## Without the dbname parameter, the driver will default to a database
### with the same name as the user. This dbname is just for instantiating a ## with the same name as the user. This dbname is just for instantiating a
### connection with the server and doesn't restrict the databases we are trying ## connection with the server and doesn't restrict the databases we are trying
### to grab metrics for. ## to grab metrics for.
### ##
address = "host=localhost user=postgres sslmode=disable" address = "host=localhost user=postgres sslmode=disable"
### A list of databases to pull metrics about. If not specified, metrics for all ## A list of databases to pull metrics about. If not specified, metrics for all
### databases are gathered. ## databases are gathered.
# databases = ["app_production", "testing"] # databases = ["app_production", "testing"]
` `

View File

@ -18,8 +18,8 @@ type Powerdns struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of sockets to gather stats about. ## An array of sockets to gather stats about.
### Specify a path to unix socket. ## Specify a path to unix socket.
unix_sockets = ["/var/run/pdns.controlsocket"] unix_sockets = ["/var/run/pdns.controlsocket"]
` `

View File

@ -30,15 +30,15 @@ func NewProcstat() *Procstat {
} }
var sampleConfig = ` var sampleConfig = `
### Must specify one of: pid_file, exe, or pattern ## Must specify one of: pid_file, exe, or pattern
### PID file to monitor process ## PID file to monitor process
pid_file = "/var/run/nginx.pid" pid_file = "/var/run/nginx.pid"
### executable name (ie, pgrep <exe>) ## executable name (ie, pgrep <exe>)
# exe = "nginx" # exe = "nginx"
### pattern as argument for pgrep (ie, pgrep -f <pattern>) ## pattern as argument for pgrep (ie, pgrep -f <pattern>)
# pattern = "nginx" # pattern = "nginx"
### Field name prefix ## Field name prefix
prefix = "" prefix = ""
` `

View File

@ -17,7 +17,7 @@ type Prometheus struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of urls to scrape metrics from. ## An array of urls to scrape metrics from.
urls = ["http://localhost:9100/metrics"] urls = ["http://localhost:9100/metrics"]
` `

View File

@ -18,7 +18,7 @@ type PuppetAgent struct {
} }
var sampleConfig = ` var sampleConfig = `
### Location of puppet last run summary file ## Location of puppet last run summary file
location = "/var/lib/puppet/state/last_run_summary.yaml" location = "/var/lib/puppet/state/last_run_summary.yaml"
` `

View File

@ -107,8 +107,8 @@ var sampleConfig = `
# username = "guest" # username = "guest"
# password = "guest" # password = "guest"
### A list of nodes to pull metrics about. If not specified, metrics for ## A list of nodes to pull metrics about. If not specified, metrics for
### all nodes are gathered. ## all nodes are gathered.
# nodes = ["rabbit@node1", "rabbit@node2"] # nodes = ["rabbit@node1", "rabbit@node2"]
` `

View File

@ -21,7 +21,7 @@ type Raindrops struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of raindrops middleware URI to gather stats. ## An array of raindrops middleware URI to gather stats.
urls = ["http://localhost:8080/_raindrops"] urls = ["http://localhost:8080/_raindrops"]
` `

View File

@ -19,14 +19,14 @@ type Redis struct {
} }
var sampleConfig = ` var sampleConfig = `
### specify servers via a url matching: ## specify servers via a url matching:
### [protocol://][:password]@address[:port] ## [protocol://][:password]@address[:port]
### e.g. ## e.g.
### tcp://localhost:6379 ## tcp://localhost:6379
### tcp://:password@192.168.99.100 ## tcp://:password@192.168.99.100
### ##
### If no servers are specified, then localhost is used as the host. ## If no servers are specified, then localhost is used as the host.
### If no port is specified, 6379 is used ## If no port is specified, 6379 is used
servers = ["tcp://localhost:6379"] servers = ["tcp://localhost:6379"]
` `

View File

@ -16,11 +16,11 @@ type RethinkDB struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of URI to gather stats about. Specify an ip or hostname ## An array of URI to gather stats about. Specify an ip or hostname
### with optional port add password. ie, ## with optional port add password. ie,
### rethinkdb://user:auth_key@10.10.3.30:28105, ## rethinkdb://user:auth_key@10.10.3.30:28105,
### rethinkdb://10.10.3.33:18832, ## rethinkdb://10.10.3.33:18832,
### 10.0.0.1:10000, etc. ## 10.0.0.1:10000, etc.
servers = ["127.0.0.1:28015"] servers = ["127.0.0.1:28015"]
` `

View File

@ -20,15 +20,15 @@ func (_ *Sensors) Description() string {
} }
var sensorsSampleConfig = ` var sensorsSampleConfig = `
### By default, telegraf gathers stats from all sensors detected by the ## By default, telegraf gathers stats from all sensors detected by the
### lm-sensors module. ## lm-sensors module.
### ##
### Only collect stats from the selected sensors. Sensors are listed as ## Only collect stats from the selected sensors. Sensors are listed as
### <chip name>:<feature name>. This information can be found by running the ## <chip name>:<feature name>. This information can be found by running the
### sensors command, e.g. sensors -u ## sensors command, e.g. sensors -u
### ##
### A * as the feature name will return all features of the chip ## A * as the feature name will return all features of the chip
### ##
# sensors = ["coretemp-isa-0000:Core 0", "coretemp-isa-0001:*"] # sensors = ["coretemp-isa-0000:Core 0", "coretemp-isa-0001:*"]
` `

View File

@ -72,11 +72,11 @@ var initNode = Node{
var NameToOid = make(map[string]string) var NameToOid = make(map[string]string)
var sampleConfig = ` var sampleConfig = `
### Use 'oids.txt' file to translate oids to names ## Use 'oids.txt' file to translate oids to names
### To generate 'oids.txt' you need to run: ## To generate 'oids.txt' you need to run:
### snmptranslate -m all -Tz -On | sed -e 's/"//g' > /tmp/oids.txt ## snmptranslate -m all -Tz -On | sed -e 's/"//g' > /tmp/oids.txt
### Or if you have an other MIB folder with custom MIBs ## Or if you have an other MIB folder with custom MIBs
### snmptranslate -M /mycustommibfolder -Tz -On -m all | sed -e 's/"//g' > oids.txt ## snmptranslate -M /mycustommibfolder -Tz -On -m all | sed -e 's/"//g' > oids.txt
snmptranslate_file = "/tmp/oids.txt" snmptranslate_file = "/tmp/oids.txt"
[[inputs.snmp.host]] [[inputs.snmp.host]]
address = "192.168.2.2:161" address = "192.168.2.2:161"

View File

@ -31,12 +31,12 @@ var queries MapQuery
var defaultServer = "Server=.;app name=telegraf;log=1;" var defaultServer = "Server=.;app name=telegraf;log=1;"
var sampleConfig = ` var sampleConfig = `
### Specify instances to monitor with a list of connection strings. ## Specify instances to monitor with a list of connection strings.
### All connection parameters are optional. ## All connection parameters are optional.
### By default, the host is localhost, listening on default port, TCP 1433. ## By default, the host is localhost, listening on default port, TCP 1433.
### for Windows, the user is the currently running AD user (SSO). ## for Windows, the user is the currently running AD user (SSO).
### See https://github.com/denisenkom/go-mssqldb for detailed connection ## See https://github.com/denisenkom/go-mssqldb for detailed connection
### parameters. ## parameters.
# servers = [ # servers = [
# "Server=192.168.1.10;Port=1433;User Id=<user>;Password=<pw>;app name=telegraf;log=1;", # "Server=192.168.1.10;Port=1433;User Id=<user>;Password=<pw>;app name=telegraf;log=1;",
# ] # ]

View File

@ -5,39 +5,39 @@
```toml ```toml
# Statsd Server # Statsd Server
[[inputs.statsd]] [[inputs.statsd]]
### Address and port to host UDP listener on ## Address and port to host UDP listener on
service_address = ":8125" service_address = ":8125"
### Delete gauges every interval (default=false) ## Delete gauges every interval (default=false)
delete_gauges = false delete_gauges = false
### Delete counters every interval (default=false) ## Delete counters every interval (default=false)
delete_counters = false delete_counters = false
### Delete sets every interval (default=false) ## Delete sets every interval (default=false)
delete_sets = false delete_sets = false
### Delete timings & histograms every interval (default=true) ## Delete timings & histograms every interval (default=true)
delete_timings = true delete_timings = true
### Percentiles to calculate for timing & histogram stats ## Percentiles to calculate for timing & histogram stats
percentiles = [90] percentiles = [90]
### convert measurement names, "." to "_" and "-" to "__" ## convert measurement names, "." to "_" and "-" to "__"
convert_names = true convert_names = true
### Statsd data translation templates, more info can be read here: ## Statsd data translation templates, more info can be read here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md#graphite ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md#graphite
# templates = [ # templates = [
# "cpu.* measurement*" # "cpu.* measurement*"
# ] # ]
### Number of UDP messages allowed to queue up, once filled, ## Number of UDP messages allowed to queue up, once filled,
### the statsd server will start dropping packets ## the statsd server will start dropping packets
allowed_pending_messages = 10000 allowed_pending_messages = 10000
### Number of timing/histogram values to track per-measurement in the ## Number of timing/histogram values to track per-measurement in the
### calculation of percentiles. Raising this limit increases the accuracy ## calculation of percentiles. Raising this limit increases the accuracy
### of percentiles but also increases the memory usage and cpu time. ## of percentiles but also increases the memory usage and cpu time.
percentile_limit = 1000 percentile_limit = 1000
### UDP packet size for the server to listen for. This will depend on the size ## UDP packet size for the server to listen for. This will depend on the size
### of the packets that the client is sending, which is usually 1500 bytes. ## of the packets that the client is sending, which is usually 1500 bytes.
udp_packet_size = 1500 udp_packet_size = 1500
``` ```

View File

@ -123,39 +123,39 @@ func (_ *Statsd) Description() string {
} }
const sampleConfig = ` const sampleConfig = `
### Address and port to host UDP listener on ## Address and port to host UDP listener on
service_address = ":8125" service_address = ":8125"
### Delete gauges every interval (default=false) ## Delete gauges every interval (default=false)
delete_gauges = false delete_gauges = false
### Delete counters every interval (default=false) ## Delete counters every interval (default=false)
delete_counters = false delete_counters = false
### Delete sets every interval (default=false) ## Delete sets every interval (default=false)
delete_sets = false delete_sets = false
### Delete timings & histograms every interval (default=true) ## Delete timings & histograms every interval (default=true)
delete_timings = true delete_timings = true
### Percentiles to calculate for timing & histogram stats ## Percentiles to calculate for timing & histogram stats
percentiles = [90] percentiles = [90]
### convert measurement names, "." to "_" and "-" to "__" ## convert measurement names, "." to "_" and "-" to "__"
convert_names = true convert_names = true
### Statsd data translation templates, more info can be read here: ## Statsd data translation templates, more info can be read here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md#graphite ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md#graphite
# templates = [ # templates = [
# "cpu.* measurement*" # "cpu.* measurement*"
# ] # ]
### Number of UDP messages allowed to queue up, once filled, ## Number of UDP messages allowed to queue up, once filled,
### the statsd server will start dropping packets ## the statsd server will start dropping packets
allowed_pending_messages = 10000 allowed_pending_messages = 10000
### Number of timing/histogram values to track per-measurement in the ## Number of timing/histogram values to track per-measurement in the
### calculation of percentiles. Raising this limit increases the accuracy ## calculation of percentiles. Raising this limit increases the accuracy
### of percentiles but also increases the memory usage and cpu time. ## of percentiles but also increases the memory usage and cpu time.
percentile_limit = 1000 percentile_limit = 1000
### UDP packet size for the server to listen for. This will depend on the size ## UDP packet size for the server to listen for. This will depend on the size
### of the packets that the client is sending, which is usually 1500 bytes. ## of the packets that the client is sending, which is usually 1500 bytes.
udp_packet_size = 1500 udp_packet_size = 1500
` `

View File

@ -28,11 +28,11 @@ func (_ *CPUStats) Description() string {
} }
var sampleConfig = ` var sampleConfig = `
### Whether to report per-cpu stats or not ## Whether to report per-cpu stats or not
percpu = true percpu = true
### Whether to report total system cpu stats or not ## Whether to report total system cpu stats or not
totalcpu = true totalcpu = true
### Comment this line if you want the raw CPU time metrics ## Comment this line if you want the raw CPU time metrics
drop = ["time_*"] drop = ["time_*"]
` `

View File

@ -21,8 +21,8 @@ func (_ *DiskStats) Description() string {
} }
var diskSampleConfig = ` var diskSampleConfig = `
### By default, telegraf gather stats for all mountpoints. ## By default, telegraf gather stats for all mountpoints.
### Setting mountpoints will restrict the stats to the specified mountpoints. ## Setting mountpoints will restrict the stats to the specified mountpoints.
# mount_points = ["/"] # mount_points = ["/"]
` `

View File

@ -21,10 +21,10 @@ func (_ *NetIOStats) Description() string {
} }
var netSampleConfig = ` var netSampleConfig = `
### By default, telegraf gathers stats from any up interface (excluding loopback) ## By default, telegraf gathers stats from any up interface (excluding loopback)
### Setting interfaces will tell it to gather these explicit interfaces, ## Setting interfaces will tell it to gather these explicit interfaces,
### regardless of status. ## regardless of status.
### ##
# interfaces = ["eth0"] # interfaces = ["eth0"]
` `

View File

@ -13,7 +13,7 @@ type Trig struct {
} }
var TrigConfig = ` var TrigConfig = `
### Set the amplitude ## Set the amplitude
amplitude = 10.0 amplitude = 10.0
` `

View File

@ -17,9 +17,9 @@ type Twemproxy struct {
} }
var sampleConfig = ` var sampleConfig = `
### Twemproxy stats address and port (no scheme) ## Twemproxy stats address and port (no scheme)
addr = "localhost:22222" addr = "localhost:22222"
### Monitor pool name ## Monitor pool name
pools = ["redis_pool", "mc_pool"] pools = ["redis_pool", "mc_pool"]
` `

View File

@ -14,12 +14,12 @@ import (
) )
var sampleConfig string = ` var sampleConfig string = `
### By default this plugin returns basic CPU and Disk statistics. ## By default this plugin returns basic CPU and Disk statistics.
### See the README file for more examples. ## See the README file for more examples.
### Uncomment examples below or write your own as you see fit. If the system ## Uncomment examples below or write your own as you see fit. If the system
### being polled for data does not have the Object at startup of the Telegraf ## being polled for data does not have the Object at startup of the Telegraf
### agent, it will not be gathered. ## agent, it will not be gathered.
### Settings: ## Settings:
# PrintValid = false # Print All matching performance counters # PrintValid = false # Print All matching performance counters
[[inputs.win_perf_counters.object]] [[inputs.win_perf_counters.object]]

View File

@ -23,15 +23,15 @@ type poolInfo struct {
} }
var sampleConfig = ` var sampleConfig = `
### ZFS kstat path ## ZFS kstat path
### If not specified, then default is: ## If not specified, then default is:
kstatPath = "/proc/spl/kstat/zfs" kstatPath = "/proc/spl/kstat/zfs"
### By default, telegraf gather all zfs stats ## By default, telegraf gather all zfs stats
### If not specified, then default is: ## If not specified, then default is:
kstatMetrics = ["arcstats", "zfetchstats", "vdev_cache_stats"] kstatMetrics = ["arcstats", "zfetchstats", "vdev_cache_stats"]
### By default, don't gather zpool stats ## By default, don't gather zpool stats
poolMetrics = false poolMetrics = false
` `

View File

@ -20,11 +20,11 @@ type Zookeeper struct {
} }
var sampleConfig = ` var sampleConfig = `
### An array of address to gather stats about. Specify an ip or hostname ## An array of address to gather stats about. Specify an ip or hostname
### with port. ie localhost:2181, 10.0.0.1:2181, etc. ## with port. ie localhost:2181, 10.0.0.1:2181, etc.
### If no servers are specified, then localhost is used as the host. ## If no servers are specified, then localhost is used as the host.
### If no port is specified, 2181 is used ## If no port is specified, 2181 is used
servers = [":2181"] servers = [":2181"]
` `

View File

@ -22,13 +22,13 @@ type Amon struct {
} }
var sampleConfig = ` var sampleConfig = `
### Amon Server Key ## Amon Server Key
server_key = "my-server-key" # required. server_key = "my-server-key" # required.
### Amon Instance URL ## Amon Instance URL
amon_instance = "https://youramoninstance" # required amon_instance = "https://youramoninstance" # required
### Connection timeout. ## Connection timeout.
# timeout = "5s" # timeout = "5s"
` `

View File

@ -52,32 +52,32 @@ const (
) )
var sampleConfig = ` var sampleConfig = `
### AMQP url ## AMQP url
url = "amqp://localhost:5672/influxdb" url = "amqp://localhost:5672/influxdb"
### AMQP exchange ## AMQP exchange
exchange = "telegraf" exchange = "telegraf"
### Telegraf tag to use as a routing key ## Telegraf tag to use as a routing key
### ie, if this tag exists, it's value will be used as the routing key ## ie, if this tag exists, it's value will be used as the routing key
routing_tag = "host" routing_tag = "host"
### InfluxDB retention policy ## InfluxDB retention policy
# retention_policy = "default" # retention_policy = "default"
### InfluxDB database ## InfluxDB database
# database = "telegraf" # database = "telegraf"
### InfluxDB precision ## InfluxDB precision
# precision = "s" # precision = "s"
### Optional SSL Config ## Optional SSL Config
# ssl_ca = "/etc/telegraf/ca.pem" # ssl_ca = "/etc/telegraf/ca.pem"
# ssl_cert = "/etc/telegraf/cert.pem" # ssl_cert = "/etc/telegraf/cert.pem"
# ssl_key = "/etc/telegraf/key.pem" # ssl_key = "/etc/telegraf/key.pem"
### Use SSL but skip chain & host verification ## Use SSL but skip chain & host verification
# insecure_skip_verify = false # insecure_skip_verify = false
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -25,10 +25,10 @@ type CloudWatch struct {
} }
var sampleConfig = ` var sampleConfig = `
### Amazon REGION ## Amazon REGION
region = 'us-east-1' region = 'us-east-1'
### Namespace for the CloudWatch MetricDatums ## Namespace for the CloudWatch MetricDatums
namespace = 'InfluxData/Telegraf' namespace = 'InfluxData/Telegraf'
` `

View File

@ -24,10 +24,10 @@ type Datadog struct {
} }
var sampleConfig = ` var sampleConfig = `
### Datadog API key ## Datadog API key
apikey = "my-secret-key" # required. apikey = "my-secret-key" # required.
### Connection timeout. ## Connection timeout.
# timeout = "5s" # timeout = "5s"
` `

View File

@ -20,13 +20,13 @@ type File struct {
} }
var sampleConfig = ` var sampleConfig = `
### Files to write to, "stdout" is a specially handled file. ## Files to write to, "stdout" is a specially handled file.
files = ["stdout", "/tmp/metrics.out"] files = ["stdout", "/tmp/metrics.out"]
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -23,11 +23,11 @@ type Graphite struct {
} }
var sampleConfig = ` var sampleConfig = `
### TCP endpoint for your graphite instance. ## TCP endpoint for your graphite instance.
servers = ["localhost:2003"] servers = ["localhost:2003"]
### Prefix metrics name ## Prefix metrics name
prefix = "" prefix = ""
### timeout in seconds for the write connection to graphite ## timeout in seconds for the write connection to graphite
timeout = 2 timeout = 2
` `

View File

@ -41,32 +41,32 @@ type InfluxDB struct {
} }
var sampleConfig = ` var sampleConfig = `
### The full HTTP or UDP endpoint URL for your InfluxDB instance. ## The full HTTP or UDP endpoint URL for your InfluxDB instance.
### Multiple urls can be specified as part of the same cluster, ## Multiple urls can be specified as part of the same cluster,
### this means that only ONE of the urls will be written to each interval. ## this means that only ONE of the urls will be written to each interval.
# urls = ["udp://localhost:8089"] # UDP endpoint example # urls = ["udp://localhost:8089"] # UDP endpoint example
urls = ["http://localhost:8086"] # required urls = ["http://localhost:8086"] # required
### The target database for metrics (telegraf will create it if not exists) ## The target database for metrics (telegraf will create it if not exists)
database = "telegraf" # required database = "telegraf" # required
### Precision of writes, valid values are "ns", "us" (or "µs"), "ms", "s", "m", "h". ## Precision of writes, valid values are "ns", "us" (or "µs"), "ms", "s", "m", "h".
### note: using "s" precision greatly improves InfluxDB compression ## note: using "s" precision greatly improves InfluxDB compression
precision = "s" precision = "s"
### Connection timeout (for the connection with InfluxDB), formatted as a string. ## Connection timeout (for the connection with InfluxDB), formatted as a string.
### If not provided, will default to 0 (no timeout) ## If not provided, will default to 0 (no timeout)
# timeout = "5s" # timeout = "5s"
# username = "telegraf" # username = "telegraf"
# password = "metricsmetricsmetricsmetrics" # password = "metricsmetricsmetricsmetrics"
### Set the user agent for HTTP POSTs (can be useful for log differentiation) ## Set the user agent for HTTP POSTs (can be useful for log differentiation)
# user_agent = "telegraf" # user_agent = "telegraf"
### Set UDP payload size, defaults to InfluxDB UDP Client default (512 bytes) ## Set UDP payload size, defaults to InfluxDB UDP Client default (512 bytes)
# udp_payload = 512 # udp_payload = 512
### Optional SSL Config ## Optional SSL Config
# ssl_ca = "/etc/telegraf/ca.pem" # ssl_ca = "/etc/telegraf/ca.pem"
# ssl_cert = "/etc/telegraf/cert.pem" # ssl_cert = "/etc/telegraf/cert.pem"
# ssl_key = "/etc/telegraf/key.pem" # ssl_key = "/etc/telegraf/key.pem"
### Use SSL but skip chain & host verification ## Use SSL but skip chain & host verification
# insecure_skip_verify = false # insecure_skip_verify = false
` `

View File

@ -45,25 +45,25 @@ type Kafka struct {
} }
var sampleConfig = ` var sampleConfig = `
### URLs of kafka brokers ## URLs of kafka brokers
brokers = ["localhost:9092"] brokers = ["localhost:9092"]
### Kafka topic for producer messages ## Kafka topic for producer messages
topic = "telegraf" topic = "telegraf"
### Telegraf tag to use as a routing key ## Telegraf tag to use as a routing key
### ie, if this tag exists, it's value will be used as the routing key ## ie, if this tag exists, it's value will be used as the routing key
routing_tag = "host" routing_tag = "host"
### Optional SSL Config ## Optional SSL Config
# ssl_ca = "/etc/telegraf/ca.pem" # ssl_ca = "/etc/telegraf/ca.pem"
# ssl_cert = "/etc/telegraf/cert.pem" # ssl_cert = "/etc/telegraf/cert.pem"
# ssl_key = "/etc/telegraf/key.pem" # ssl_key = "/etc/telegraf/key.pem"
### Use SSL but skip chain & host verification ## Use SSL but skip chain & host verification
# insecure_skip_verify = false # insecure_skip_verify = false
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -28,16 +28,16 @@ type KinesisOutput struct {
} }
var sampleConfig = ` var sampleConfig = `
### Amazon REGION of kinesis endpoint. ## Amazon REGION of kinesis endpoint.
region = "ap-southeast-2" region = "ap-southeast-2"
### Kinesis StreamName must exist prior to starting telegraf. ## Kinesis StreamName must exist prior to starting telegraf.
streamname = "StreamName" streamname = "StreamName"
### PartitionKey as used for sharding data. ## PartitionKey as used for sharding data.
partitionkey = "PartitionKey" partitionkey = "PartitionKey"
### format of the Data payload in the kinesis PutRecord, supported ## format of the Data payload in the kinesis PutRecord, supported
### String and Custom. ## String and Custom.
format = "string" format = "string"
### debug will show upstream aws messages. ## debug will show upstream aws messages.
debug = false debug = false
` `

View File

@ -23,20 +23,20 @@ type Librato struct {
} }
var sampleConfig = ` var sampleConfig = `
### Librator API Docs ## Librator API Docs
### http://dev.librato.com/v1/metrics-authentication ## http://dev.librato.com/v1/metrics-authentication
### Librato API user ## Librato API user
api_user = "telegraf@influxdb.com" # required. api_user = "telegraf@influxdb.com" # required.
### Librato API token ## Librato API token
api_token = "my-secret-token" # required. api_token = "my-secret-token" # required.
### Tag Field to populate source attribute (optional) ## Tag Field to populate source attribute (optional)
### This is typically the _hostname_ from which the metric was obtained. ## This is typically the _hostname_ from which the metric was obtained.
source_tag = "hostname" source_tag = "hostname"
### Connection timeout. ## Connection timeout.
# timeout = "5s" # timeout = "5s"
` `

View File

@ -16,26 +16,26 @@ import (
var sampleConfig = ` var sampleConfig = `
servers = ["localhost:1883"] # required. servers = ["localhost:1883"] # required.
### MQTT outputs send metrics to this topic format ## MQTT outputs send metrics to this topic format
### "<topic_prefix>/<hostname>/<pluginname>/" ## "<topic_prefix>/<hostname>/<pluginname>/"
### ex: prefix/web01.example.com/mem ## ex: prefix/web01.example.com/mem
topic_prefix = "telegraf" topic_prefix = "telegraf"
### username and password to connect MQTT server. ## username and password to connect MQTT server.
# username = "telegraf" # username = "telegraf"
# password = "metricsmetricsmetricsmetrics" # password = "metricsmetricsmetricsmetrics"
### Optional SSL Config ## Optional SSL Config
# ssl_ca = "/etc/telegraf/ca.pem" # ssl_ca = "/etc/telegraf/ca.pem"
# ssl_cert = "/etc/telegraf/cert.pem" # ssl_cert = "/etc/telegraf/cert.pem"
# ssl_key = "/etc/telegraf/key.pem" # ssl_key = "/etc/telegraf/key.pem"
### Use SSL but skip chain & host verification ## Use SSL but skip chain & host verification
# insecure_skip_verify = false # insecure_skip_verify = false
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -19,15 +19,15 @@ type NSQ struct {
} }
var sampleConfig = ` var sampleConfig = `
### Location of nsqd instance listening on TCP ## Location of nsqd instance listening on TCP
server = "localhost:4150" server = "localhost:4150"
### NSQ topic for producer messages ## NSQ topic for producer messages
topic = "telegraf" topic = "telegraf"
### Data format to output. This can be "influx" or "graphite" ## Data format to output. This can be "influx" or "graphite"
### Each data format has it's own unique set of configuration options, read ## Each data format has it's own unique set of configuration options, read
### more about them here: ## more about them here:
### https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md ## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_OUTPUT.md
data_format = "influx" data_format = "influx"
` `

View File

@ -22,17 +22,17 @@ type OpenTSDB struct {
} }
var sampleConfig = ` var sampleConfig = `
### prefix for metrics keys ## prefix for metrics keys
prefix = "my.specific.prefix." prefix = "my.specific.prefix."
## Telnet Mode ## ## Telnet Mode ##
### DNS name of the OpenTSDB server in telnet mode ## DNS name of the OpenTSDB server in telnet mode
host = "opentsdb.example.com" host = "opentsdb.example.com"
### Port of the OpenTSDB server in telnet mode ## Port of the OpenTSDB server in telnet mode
port = 4242 port = 4242
### Debug true - Prints OpenTSDB communication ## Debug true - Prints OpenTSDB communication
debug = false debug = false
` `

View File

@ -16,7 +16,7 @@ type PrometheusClient struct {
} }
var sampleConfig = ` var sampleConfig = `
### Address to listen on ## Address to listen on
# listen = ":9126" # listen = ":9126"
` `

View File

@ -21,11 +21,11 @@ type Riemann struct {
} }
var sampleConfig = ` var sampleConfig = `
### URL of server ## URL of server
url = "localhost:5555" url = "localhost:5555"
### transport protocol to use either tcp or udp ## transport protocol to use either tcp or udp
transport = "tcp" transport = "tcp"
### separator to use between input name and field name in Riemann service name ## separator to use between input name and field name in Riemann service name
separator = " " separator = " "
` `