Kafka output producer, send telegraf metrics to Kafka brokers
Closes #38
This commit is contained in:
parent
434267898b
commit
d1f965ae30
|
@ -1,6 +1,7 @@
|
||||||
## v0.1.7 [unreleased]
|
## v0.1.7 [unreleased]
|
||||||
|
|
||||||
### Features
|
### Features
|
||||||
|
- [#38](https://github.com/influxdb/telegraf/pull/38): Kafka output sink.
|
||||||
- [#133](https://github.com/influxdb/telegraf/pull/133): Add plugin.Gather error logging. Thanks @nickscript0!
|
- [#133](https://github.com/influxdb/telegraf/pull/133): Add plugin.Gather error logging. Thanks @nickscript0!
|
||||||
- [#136](https://github.com/influxdb/telegraf/issues/136): Add a -usage flag for printing usage of a single plugin.
|
- [#136](https://github.com/influxdb/telegraf/issues/136): Add a -usage flag for printing usage of a single plugin.
|
||||||
- [#137](https://github.com/influxdb/telegraf/issues/137): Memcached: fix when a value contains a space
|
- [#137](https://github.com/influxdb/telegraf/issues/137): Memcached: fix when a value contains a space
|
||||||
|
|
11
README.md
11
README.md
|
@ -103,17 +103,18 @@ at 192.168.59.103:8086, tagging measurements with dc="denver-1". It will output
|
||||||
measurements at a 10s interval and will collect totalcpu & percpu data.
|
measurements at a 10s interval and will collect totalcpu & percpu data.
|
||||||
|
|
||||||
```
|
```
|
||||||
[outputs]
|
|
||||||
[outputs.influxdb]
|
|
||||||
url = "http://192.168.59.103:8086" # required.
|
|
||||||
database = "telegraf" # required.
|
|
||||||
|
|
||||||
[tags]
|
[tags]
|
||||||
dc = "denver-1"
|
dc = "denver-1"
|
||||||
|
|
||||||
[agent]
|
[agent]
|
||||||
interval = "10s"
|
interval = "10s"
|
||||||
|
|
||||||
|
# OUTPUTS
|
||||||
|
[outputs]
|
||||||
|
[outputs.influxdb]
|
||||||
|
url = "http://192.168.59.103:8086" # required.
|
||||||
|
database = "telegraf" # required.
|
||||||
|
|
||||||
# PLUGINS
|
# PLUGINS
|
||||||
[cpu]
|
[cpu]
|
||||||
percpu = true
|
percpu = true
|
||||||
|
|
8
agent.go
8
agent.go
|
@ -74,6 +74,9 @@ func (a *Agent) Connect() error {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
if a.Debug {
|
||||||
|
log.Printf("Successfully connected to output: %s\n", o.name)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -160,6 +163,8 @@ func (a *Agent) LoadPlugins(pluginsFilter string) ([]string, error) {
|
||||||
return names, nil
|
return names, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// crankParallel runs the plugins that are using the same reporting interval
|
||||||
|
// as the telegraf agent.
|
||||||
func (a *Agent) crankParallel() error {
|
func (a *Agent) crankParallel() error {
|
||||||
points := make(chan *BatchPoints, len(a.plugins))
|
points := make(chan *BatchPoints, len(a.plugins))
|
||||||
|
|
||||||
|
@ -203,6 +208,7 @@ func (a *Agent) crankParallel() error {
|
||||||
return a.flush(&bp)
|
return a.flush(&bp)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// crank is mostly for test purposes.
|
||||||
func (a *Agent) crank() error {
|
func (a *Agent) crank() error {
|
||||||
var bp BatchPoints
|
var bp BatchPoints
|
||||||
|
|
||||||
|
@ -223,6 +229,8 @@ func (a *Agent) crank() error {
|
||||||
return a.flush(&bp)
|
return a.flush(&bp)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// crankSeparate runs the plugins that have been configured with their own
|
||||||
|
// reporting interval.
|
||||||
func (a *Agent) crankSeparate(shutdown chan struct{}, plugin *runningPlugin) error {
|
func (a *Agent) crankSeparate(shutdown chan struct{}, plugin *runningPlugin) error {
|
||||||
ticker := time.NewTicker(plugin.config.Interval)
|
ticker := time.NewTicker(plugin.config.Interval)
|
||||||
|
|
||||||
|
|
|
@ -114,11 +114,8 @@ func main() {
|
||||||
}
|
}
|
||||||
|
|
||||||
shutdown := make(chan struct{})
|
shutdown := make(chan struct{})
|
||||||
|
|
||||||
signals := make(chan os.Signal)
|
signals := make(chan os.Signal)
|
||||||
|
|
||||||
signal.Notify(signals, os.Interrupt)
|
signal.Notify(signals, os.Interrupt)
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
<-signals
|
<-signals
|
||||||
close(shutdown)
|
close(shutdown)
|
||||||
|
|
|
@ -356,6 +356,7 @@ var header = `# Telegraf configuration
|
||||||
# debug = false
|
# debug = false
|
||||||
# hostname = "prod3241"
|
# hostname = "prod3241"
|
||||||
|
|
||||||
|
|
||||||
###############################################################################
|
###############################################################################
|
||||||
# OUTPUTS #
|
# OUTPUTS #
|
||||||
###############################################################################
|
###############################################################################
|
||||||
|
@ -368,7 +369,6 @@ var header2 = `
|
||||||
###############################################################################
|
###############################################################################
|
||||||
# PLUGINS #
|
# PLUGINS #
|
||||||
###############################################################################
|
###############################################################################
|
||||||
|
|
||||||
`
|
`
|
||||||
|
|
||||||
// PrintSampleConfig prints the sample config!
|
// PrintSampleConfig prints the sample config!
|
||||||
|
|
|
@ -3,4 +3,5 @@ package all
|
||||||
import (
|
import (
|
||||||
_ "github.com/influxdb/telegraf/outputs/datadog"
|
_ "github.com/influxdb/telegraf/outputs/datadog"
|
||||||
_ "github.com/influxdb/telegraf/outputs/influxdb"
|
_ "github.com/influxdb/telegraf/outputs/influxdb"
|
||||||
|
_ "github.com/influxdb/telegraf/outputs/kafka"
|
||||||
)
|
)
|
||||||
|
|
|
@ -9,6 +9,8 @@ import (
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/influxdb/telegraf/testutil"
|
||||||
|
|
||||||
"github.com/influxdb/influxdb/client"
|
"github.com/influxdb/influxdb/client"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
@ -25,18 +27,6 @@ func fakeDatadog() *Datadog {
|
||||||
return d
|
return d
|
||||||
}
|
}
|
||||||
|
|
||||||
func testData() client.BatchPoints {
|
|
||||||
var bp client.BatchPoints
|
|
||||||
bp.Time = time.Now()
|
|
||||||
bp.Tags = map[string]string{"tag1": "value1"}
|
|
||||||
bp.Points = []client.Point{
|
|
||||||
{
|
|
||||||
Fields: map[string]interface{}{"value": 1.0},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
return bp
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestUriOverride(t *testing.T) {
|
func TestUriOverride(t *testing.T) {
|
||||||
ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
ts := httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||||
w.WriteHeader(http.StatusOK)
|
w.WriteHeader(http.StatusOK)
|
||||||
|
@ -48,7 +38,7 @@ func TestUriOverride(t *testing.T) {
|
||||||
d.Apikey = "123456"
|
d.Apikey = "123456"
|
||||||
err := d.Connect()
|
err := d.Connect()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
err = d.Write(testData())
|
err = d.Write(testutil.MockBatchPoints())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -67,7 +57,7 @@ func TestBadStatusCode(t *testing.T) {
|
||||||
d.Apikey = "123456"
|
d.Apikey = "123456"
|
||||||
err := d.Connect()
|
err := d.Connect()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
err = d.Write(testData())
|
err = d.Write(testutil.MockBatchPoints())
|
||||||
if err == nil {
|
if err == nil {
|
||||||
t.Errorf("error expected but none returned")
|
t.Errorf("error expected but none returned")
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -0,0 +1,91 @@
|
||||||
|
package kafka
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
|
||||||
|
"github.com/Shopify/sarama"
|
||||||
|
"github.com/influxdb/influxdb/client"
|
||||||
|
"github.com/influxdb/telegraf/outputs"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Kafka struct {
|
||||||
|
// Kafka brokers to send metrics to
|
||||||
|
Brokers []string
|
||||||
|
// Kafka topic
|
||||||
|
Topic string
|
||||||
|
|
||||||
|
producer sarama.SyncProducer
|
||||||
|
}
|
||||||
|
|
||||||
|
var sampleConfig = `
|
||||||
|
# URLs of kafka brokers
|
||||||
|
brokers = ["localhost:9092"]
|
||||||
|
# Kafka topic for producer messages
|
||||||
|
topic = "telegraf"
|
||||||
|
`
|
||||||
|
|
||||||
|
func (k *Kafka) Connect() error {
|
||||||
|
producer, err := sarama.NewSyncProducer(k.Brokers, nil)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
k.producer = producer
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (k *Kafka) Close() error {
|
||||||
|
return k.producer.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (k *Kafka) SampleConfig() string {
|
||||||
|
return sampleConfig
|
||||||
|
}
|
||||||
|
|
||||||
|
func (k *Kafka) Description() string {
|
||||||
|
return "Configuration for the Kafka server to send metrics to"
|
||||||
|
}
|
||||||
|
|
||||||
|
func (k *Kafka) Write(bp client.BatchPoints) error {
|
||||||
|
if len(bp.Points) == 0 {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, p := range bp.Points {
|
||||||
|
// Combine tags from Point and BatchPoints and grab the resulting
|
||||||
|
// line-protocol output string to write to Kafka
|
||||||
|
var value string
|
||||||
|
if p.Raw != "" {
|
||||||
|
value = p.Raw
|
||||||
|
} else {
|
||||||
|
for k, v := range bp.Tags {
|
||||||
|
if p.Tags == nil {
|
||||||
|
p.Tags = make(map[string]string, len(bp.Tags))
|
||||||
|
}
|
||||||
|
p.Tags[k] = v
|
||||||
|
}
|
||||||
|
value = p.MarshalString()
|
||||||
|
}
|
||||||
|
|
||||||
|
m := &sarama.ProducerMessage{
|
||||||
|
Topic: k.Topic,
|
||||||
|
Value: sarama.StringEncoder(value),
|
||||||
|
}
|
||||||
|
if h, ok := p.Tags["host"]; ok {
|
||||||
|
m.Key = sarama.StringEncoder(h)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, _, err := k.producer.SendMessage(m)
|
||||||
|
if err != nil {
|
||||||
|
return errors.New(fmt.Sprintf("FAILED to send kafka message: %s\n",
|
||||||
|
err))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
outputs.Add("kafka", func() outputs.Output {
|
||||||
|
return &Kafka{}
|
||||||
|
})
|
||||||
|
}
|
|
@ -0,0 +1,28 @@
|
||||||
|
package kafka
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/influxdb/telegraf/testutil"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestConnectAndWrite(t *testing.T) {
|
||||||
|
if testing.Short() {
|
||||||
|
t.Skip("Skipping integration test in short mode")
|
||||||
|
}
|
||||||
|
|
||||||
|
brokers := []string{testutil.GetLocalHost() + ":9092"}
|
||||||
|
k := &Kafka{
|
||||||
|
Brokers: brokers,
|
||||||
|
Topic: "Test",
|
||||||
|
}
|
||||||
|
|
||||||
|
// Verify that we can connect to the Kafka broker
|
||||||
|
err := k.Connect()
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// Verify that we can successfully write data to the kafka broker
|
||||||
|
err = k.Write(testutil.MockBatchPoints())
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
|
@ -4,6 +4,9 @@ import (
|
||||||
"net"
|
"net"
|
||||||
"net/url"
|
"net/url"
|
||||||
"os"
|
"os"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/influxdb/influxdb/client"
|
||||||
)
|
)
|
||||||
|
|
||||||
var localhost = "localhost"
|
var localhost = "localhost"
|
||||||
|
@ -27,3 +30,17 @@ func GetLocalHost() string {
|
||||||
}
|
}
|
||||||
return localhost
|
return localhost
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// MockBatchPoints returns a mock BatchPoints object for using in unit tests
|
||||||
|
// of telegraf output sinks.
|
||||||
|
func MockBatchPoints() client.BatchPoints {
|
||||||
|
var bp client.BatchPoints
|
||||||
|
bp.Time = time.Now()
|
||||||
|
bp.Tags = map[string]string{"tag1": "value1"}
|
||||||
|
bp.Points = []client.Point{
|
||||||
|
{
|
||||||
|
Fields: map[string]interface{}{"value": 1.0},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
return bp
|
||||||
|
}
|
||||||
|
|
Loading…
Reference in New Issue