Commit Graph

39 Commits

Author SHA1 Message Date
Daniel Nelson f6b302621e
Do not add invalid timestamps to kafka messages (#6908) 2020-01-15 15:26:50 -08:00
Daniel Nelson 2486006495
Add kafka SASL version control to kafka_consumer (#6350) 2020-01-02 16:27:26 -08:00
Daniel Nelson 317c823bfc
Set message timestamp to the metric time in kafka output (#6746) 2019-12-03 11:48:53 -08:00
Daniel Nelson 6cbaf890d9
Use github.com/gofrs/uuid 2.1.0 (#6636) 2019-11-07 17:39:19 -08:00
Daniel Nelson c9107015b0
Emit sarama library log messages at debug level (#6091) 2019-07-11 13:50:12 -07:00
Daniel Nelson 4197426a73
Ignore errors serializing single metrics (#5943) 2019-06-03 17:34:48 -07:00
Daniel Nelson d2cf9a7157
Add support for static and random routing keys in kafka output (#4579) 2018-08-21 12:44:10 -07:00
Daniel Nelson 886d8cc840
Drop message batches in kafka output if too large (#4565) 2018-08-17 13:51:21 -07:00
Daniel Nelson 34614582a7
Use snake case in kafka output config 2018-08-15 14:12:22 -07:00
Mauro Murari 61513c64b8
Add message 'max_bytes' configuration (#4537) 2018-08-15 14:05:26 -07:00
Daniel Nelson f4032fc78d Add support for lz4 compression to kafka output (#4492) 2018-07-31 16:09:30 -06:00
Daniel Nelson 943dcc0c49 Send all messages before waiting for results in kafka output (#4491) 2018-07-31 16:08:04 -06:00
Daniel Nelson 49a5dea536
Update client_id in kafka input and output readme 2018-07-13 14:00:25 -07:00
Rion c8f00030dd Add support for setting kafka client id (#4418) 2018-07-13 13:53:56 -07:00
Daniel Nelson de355b76d6
Simplify testing with TLS (#4095) 2018-05-04 16:33:23 -07:00
Daniel Nelson b2bb44363a
Update kafka readme 2018-05-04 14:39:31 -07:00
Daniel Nelson fd63591b15
Fix grammar 2018-05-03 17:26:01 -07:00
Daniel Nelson 2108582b43
Clarify max_retry option in kafka output 2018-05-03 17:22:49 -07:00
Daniel Nelson 402a0f16e1
Fix typo 2017-09-06 14:19:42 -07:00
Pavel Gurkov 5d4eec606f Add Kafka output plugin topic_suffix option (#3196) 2017-09-06 14:18:26 -07:00
Seuf cd103c85db Added SASL options for ouput kafka plugin (#2721) 2017-04-27 11:50:25 -07:00
Damien Krotkine 153304d92b it's -> its (#2728) 2017-04-27 11:10:00 -07:00
Damien Krotkine cb9aecbf04 it's -> its (#2729) 2017-04-27 11:06:40 -07:00
Cameron Sparr 285be648c4
Godeps update
closes #2356
2017-02-02 09:52:06 +00:00
Cameron Sparr a093ec1eaa Kafka output fixup 2016-12-06 15:38:59 +00:00
Jack Zampolin b0ef506a88 Add Kafka output readme (#1609) 2016-08-08 23:10:07 +01:00
Cameron Sparr a30b1a394f Kafka output: set max_retry=3 & required_acks=-1 as defaults
closes #1113
2016-04-29 18:51:45 -06:00
Cameron Sparr e436b2d720 Cleanup & standardize config file
changes:

- -sample-config will now comment out all but a few default plugins.
- config file parse errors will output path to bad conf file.
- cleanup 80-char line-length and some other style issues.
- default package conf file will now have all plugins, but commented
  out.

closes #199
closes #944
2016-04-01 10:59:53 -06:00
Florent Ramière 8c3371c4ac Use numerical codes instead of symbolic ones 2016-04-01 10:08:55 -06:00
Florent Ramière 6ff0fc6d83 Add compression/acks/retry conf to Kafka output plugin
The following configuration is now possible

	## CompressionCodec represents the various compression codecs
recognized by Kafka in messages.
	##  "none" : No compression
	##  "gzip" : Gzip compression
	##  "snappy" : Snappy compression
	# compression_codec = "none"

	##  RequiredAcks is used in Produce Requests to tell the broker how
many replica acknowledgements it must see before responding
	##  "none" : the producer never waits for an acknowledgement from the
broker. This option provides the lowest latency but the weakest
durability guarantees (some data will be lost when a server fails).
	##  "leader" : the producer gets an acknowledgement after the leader
replica has received the data. This option provides better durability
as the client waits until the server acknowledges the request as
successful (only messages that were written to the now-dead leader but
not yet replicated will be lost).
	##  "leader_and_replicas" : the producer gets an acknowledgement after
all in-sync replicas have received the data. This option provides the
best durability, we guarantee that no messages will be lost as long as
at least one in sync replica remains.
	# required_acks = "leader_and_replicas"

	##  The total number of times to retry sending a message
	# max_retry = "3"
2016-04-01 10:08:55 -06:00
Cameron Sparr 8d2e5f0bda Seems to be a toml parse bug around triple pounds 2016-02-18 14:36:03 -07:00
Cameron Sparr 7def6663bd Root directory cleanup 2016-02-18 13:37:36 -07:00
Cameron Sparr a9c135488e Add Serializer plugins, and 'file' output plugin 2016-02-12 14:13:49 -07:00
Cameron Sparr bd9c5b6995 mqtt output: cleanup, implement TLS
Also normalize TLS config across all output plugins and normalize
comment strings as well.
2016-02-04 10:44:37 -07:00
Cameron Sparr c549ab907a Throughout telegraf, use telegraf.Metric rather than client.Point
closes #599
2016-01-27 23:47:32 -07:00
Cameron Sparr 9c0d14bb60 Create public models for telegraf metrics, accumlator, plugins
This will basically make the root directory a place for storing the
major telegraf interfaces, which will make telegraf's godoc looks quite
a bit nicer. And make it easier for contributors to lookup the few data
types that they actually care about.

closes #564
2016-01-27 15:42:50 -07:00
Jack Zampolin 0cdf1b07e9 Fix issue 524 2016-01-20 10:57:35 -08:00
Hannu Valtonen c313af1b24 kafka: Add support for using TLS authentication for the kafka output
With the advent of Kafka 0.9.0+ it is possible to set up TLS client
certificate based authentication to limit access to Kafka.

Four new configuration variables are specified for setting up the
authentication. If they're not set the behavior stays the same as
before the change.

closes #541
2016-01-18 11:17:01 -07:00
Cameron Sparr 9c5db1057d renaming plugins -> inputs 2016-01-07 15:04:30 -07:00