2016-01-18 19:39:14 +00:00
|
|
|
package cloudwatch
|
|
|
|
|
|
|
|
import (
|
|
|
|
"log"
|
|
|
|
"math"
|
|
|
|
"sort"
|
|
|
|
"strings"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/aws/aws-sdk-go/aws"
|
|
|
|
"github.com/aws/aws-sdk-go/service/cloudwatch"
|
|
|
|
|
2016-01-27 21:21:36 +00:00
|
|
|
"github.com/influxdata/telegraf"
|
2016-05-25 11:30:39 +00:00
|
|
|
internalaws "github.com/influxdata/telegraf/internal/config/aws"
|
2016-01-27 23:15:14 +00:00
|
|
|
"github.com/influxdata/telegraf/plugins/outputs"
|
2016-01-18 19:39:14 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
type CloudWatch struct {
|
2018-07-31 22:07:21 +00:00
|
|
|
Region string `toml:"region"`
|
|
|
|
AccessKey string `toml:"access_key"`
|
|
|
|
SecretKey string `toml:"secret_key"`
|
|
|
|
RoleARN string `toml:"role_arn"`
|
|
|
|
Profile string `toml:"profile"`
|
|
|
|
Filename string `toml:"shared_credential_file"`
|
|
|
|
Token string `toml:"token"`
|
|
|
|
EndpointURL string `toml:"endpoint_url"`
|
2016-05-25 11:30:39 +00:00
|
|
|
|
|
|
|
Namespace string `toml:"namespace"` // CloudWatch Metrics Namespace
|
2016-01-18 19:39:14 +00:00
|
|
|
svc *cloudwatch.CloudWatch
|
2018-08-01 22:43:17 +00:00
|
|
|
|
|
|
|
WriteStatistics bool `toml:"write_statistics"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type statisticType int
|
|
|
|
|
|
|
|
const (
|
|
|
|
statisticTypeNone statisticType = iota
|
|
|
|
statisticTypeMax
|
|
|
|
statisticTypeMin
|
|
|
|
statisticTypeSum
|
|
|
|
statisticTypeCount
|
|
|
|
)
|
|
|
|
|
|
|
|
type cloudwatchField interface {
|
|
|
|
addValue(sType statisticType, value float64)
|
|
|
|
buildDatum() []*cloudwatch.MetricDatum
|
|
|
|
}
|
|
|
|
|
|
|
|
type statisticField struct {
|
|
|
|
metricName string
|
|
|
|
fieldName string
|
|
|
|
tags map[string]string
|
|
|
|
values map[statisticType]float64
|
|
|
|
timestamp time.Time
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *statisticField) addValue(sType statisticType, value float64) {
|
|
|
|
if sType != statisticTypeNone {
|
|
|
|
f.values[sType] = value
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *statisticField) buildDatum() []*cloudwatch.MetricDatum {
|
|
|
|
|
|
|
|
var datums []*cloudwatch.MetricDatum
|
|
|
|
|
|
|
|
if f.hasAllFields() {
|
|
|
|
// If we have all required fields, we build datum with StatisticValues
|
|
|
|
min, _ := f.values[statisticTypeMin]
|
|
|
|
max, _ := f.values[statisticTypeMax]
|
|
|
|
sum, _ := f.values[statisticTypeSum]
|
|
|
|
count, _ := f.values[statisticTypeCount]
|
|
|
|
|
|
|
|
datum := &cloudwatch.MetricDatum{
|
|
|
|
MetricName: aws.String(strings.Join([]string{f.metricName, f.fieldName}, "_")),
|
|
|
|
Dimensions: BuildDimensions(f.tags),
|
|
|
|
Timestamp: aws.Time(f.timestamp),
|
|
|
|
StatisticValues: &cloudwatch.StatisticSet{
|
|
|
|
Minimum: aws.Float64(min),
|
|
|
|
Maximum: aws.Float64(max),
|
|
|
|
Sum: aws.Float64(sum),
|
|
|
|
SampleCount: aws.Float64(count),
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
datums = append(datums, datum)
|
|
|
|
|
|
|
|
} else {
|
|
|
|
// If we don't have all required fields, we build each field as independent datum
|
|
|
|
for sType, value := range f.values {
|
|
|
|
datum := &cloudwatch.MetricDatum{
|
|
|
|
Value: aws.Float64(value),
|
|
|
|
Dimensions: BuildDimensions(f.tags),
|
|
|
|
Timestamp: aws.Time(f.timestamp),
|
|
|
|
}
|
|
|
|
|
|
|
|
switch sType {
|
|
|
|
case statisticTypeMin:
|
|
|
|
datum.MetricName = aws.String(strings.Join([]string{f.metricName, f.fieldName, "min"}, "_"))
|
|
|
|
case statisticTypeMax:
|
|
|
|
datum.MetricName = aws.String(strings.Join([]string{f.metricName, f.fieldName, "max"}, "_"))
|
|
|
|
case statisticTypeSum:
|
|
|
|
datum.MetricName = aws.String(strings.Join([]string{f.metricName, f.fieldName, "sum"}, "_"))
|
|
|
|
case statisticTypeCount:
|
|
|
|
datum.MetricName = aws.String(strings.Join([]string{f.metricName, f.fieldName, "count"}, "_"))
|
|
|
|
default:
|
|
|
|
// should not be here
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
datums = append(datums, datum)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return datums
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *statisticField) hasAllFields() bool {
|
|
|
|
|
|
|
|
_, hasMin := f.values[statisticTypeMin]
|
|
|
|
_, hasMax := f.values[statisticTypeMax]
|
|
|
|
_, hasSum := f.values[statisticTypeSum]
|
|
|
|
_, hasCount := f.values[statisticTypeCount]
|
|
|
|
|
|
|
|
return hasMin && hasMax && hasSum && hasCount
|
|
|
|
}
|
|
|
|
|
|
|
|
type valueField struct {
|
|
|
|
metricName string
|
|
|
|
fieldName string
|
|
|
|
tags map[string]string
|
|
|
|
value float64
|
|
|
|
timestamp time.Time
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *valueField) addValue(sType statisticType, value float64) {
|
|
|
|
if sType == statisticTypeNone {
|
|
|
|
f.value = value
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (f *valueField) buildDatum() []*cloudwatch.MetricDatum {
|
|
|
|
|
|
|
|
return []*cloudwatch.MetricDatum{
|
|
|
|
{
|
|
|
|
MetricName: aws.String(strings.Join([]string{f.metricName, f.fieldName}, "_")),
|
|
|
|
Value: aws.Float64(f.value),
|
|
|
|
Dimensions: BuildDimensions(f.tags),
|
|
|
|
Timestamp: aws.Time(f.timestamp),
|
|
|
|
},
|
|
|
|
}
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
var sampleConfig = `
|
2016-02-18 21:26:51 +00:00
|
|
|
## Amazon REGION
|
2016-11-04 13:16:41 +00:00
|
|
|
region = "us-east-1"
|
2016-01-18 19:39:14 +00:00
|
|
|
|
2016-04-23 18:19:04 +00:00
|
|
|
## Amazon Credentials
|
|
|
|
## Credentials are loaded in the following order
|
2016-05-25 11:30:39 +00:00
|
|
|
## 1) Assumed credentials via STS if role_arn is specified
|
|
|
|
## 2) explicit credentials from 'access_key' and 'secret_key'
|
|
|
|
## 3) shared profile from 'profile'
|
|
|
|
## 4) environment variables
|
|
|
|
## 5) shared credentials file
|
|
|
|
## 6) EC2 Instance Profile
|
2016-04-23 18:19:04 +00:00
|
|
|
#access_key = ""
|
|
|
|
#secret_key = ""
|
2016-05-25 11:30:39 +00:00
|
|
|
#token = ""
|
|
|
|
#role_arn = ""
|
|
|
|
#profile = ""
|
|
|
|
#shared_credential_file = ""
|
2016-04-23 18:19:04 +00:00
|
|
|
|
2018-07-31 22:07:21 +00:00
|
|
|
## Endpoint to make request against, the correct endpoint is automatically
|
|
|
|
## determined and this option should only be set if you wish to override the
|
|
|
|
## default.
|
|
|
|
## ex: endpoint_url = "http://localhost:8000"
|
|
|
|
# endpoint_url = ""
|
|
|
|
|
2016-02-18 21:26:51 +00:00
|
|
|
## Namespace for the CloudWatch MetricDatums
|
2016-11-04 13:16:41 +00:00
|
|
|
namespace = "InfluxData/Telegraf"
|
2018-08-01 22:43:17 +00:00
|
|
|
|
|
|
|
## If you have a large amount of metrics, you should consider to send statistic
|
|
|
|
## values instead of raw metrics which could not only improve performance but
|
|
|
|
## also save AWS API cost. If enable this flag, this plugin would parse the required
|
|
|
|
## CloudWatch statistic fields (count, min, max, and sum) and send them to CloudWatch.
|
|
|
|
## You could use basicstats aggregator to calculate those fields. If not all statistic
|
|
|
|
## fields are available, all fields would still be sent as raw metrics.
|
|
|
|
# write_statistics = false
|
2016-01-18 19:39:14 +00:00
|
|
|
`
|
|
|
|
|
|
|
|
func (c *CloudWatch) SampleConfig() string {
|
|
|
|
return sampleConfig
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *CloudWatch) Description() string {
|
|
|
|
return "Configuration for AWS CloudWatch output."
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *CloudWatch) Connect() error {
|
2016-05-25 11:30:39 +00:00
|
|
|
credentialConfig := &internalaws.CredentialConfig{
|
2018-07-31 22:07:21 +00:00
|
|
|
Region: c.Region,
|
|
|
|
AccessKey: c.AccessKey,
|
|
|
|
SecretKey: c.SecretKey,
|
|
|
|
RoleARN: c.RoleARN,
|
|
|
|
Profile: c.Profile,
|
|
|
|
Filename: c.Filename,
|
|
|
|
Token: c.Token,
|
|
|
|
EndpointURL: c.EndpointURL,
|
2016-04-23 18:19:04 +00:00
|
|
|
}
|
2016-05-25 11:30:39 +00:00
|
|
|
configProvider := credentialConfig.Credentials()
|
2017-10-13 18:04:40 +00:00
|
|
|
c.svc = cloudwatch.New(configProvider)
|
|
|
|
return nil
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (c *CloudWatch) Close() error {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2016-01-27 23:15:14 +00:00
|
|
|
func (c *CloudWatch) Write(metrics []telegraf.Metric) error {
|
2018-07-23 18:00:35 +00:00
|
|
|
|
|
|
|
var datums []*cloudwatch.MetricDatum
|
2016-01-27 23:15:14 +00:00
|
|
|
for _, m := range metrics {
|
2018-08-01 22:43:17 +00:00
|
|
|
d := BuildMetricDatum(c.WriteStatistics, m)
|
2018-07-23 18:00:35 +00:00
|
|
|
datums = append(datums, d...)
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
|
|
|
|
2016-01-27 23:15:14 +00:00
|
|
|
const maxDatumsPerCall = 20 // PutMetricData only supports up to 20 data metrics per call
|
2016-01-18 19:39:14 +00:00
|
|
|
|
|
|
|
for _, partition := range PartitionDatums(maxDatumsPerCall, datums) {
|
|
|
|
err := c.WriteToCloudWatch(partition)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (c *CloudWatch) WriteToCloudWatch(datums []*cloudwatch.MetricDatum) error {
|
|
|
|
params := &cloudwatch.PutMetricDataInput{
|
|
|
|
MetricData: datums,
|
|
|
|
Namespace: aws.String(c.Namespace),
|
|
|
|
}
|
|
|
|
|
|
|
|
_, err := c.svc.PutMetricData(params)
|
|
|
|
|
|
|
|
if err != nil {
|
2016-09-30 21:37:56 +00:00
|
|
|
log.Printf("E! CloudWatch: Unable to write to CloudWatch : %+v \n", err.Error())
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Partition the MetricDatums into smaller slices of a max size so that are under the limit
|
|
|
|
// for the AWS API calls.
|
|
|
|
func PartitionDatums(size int, datums []*cloudwatch.MetricDatum) [][]*cloudwatch.MetricDatum {
|
|
|
|
|
|
|
|
numberOfPartitions := len(datums) / size
|
|
|
|
if len(datums)%size != 0 {
|
|
|
|
numberOfPartitions += 1
|
|
|
|
}
|
|
|
|
|
|
|
|
partitions := make([][]*cloudwatch.MetricDatum, numberOfPartitions)
|
|
|
|
|
|
|
|
for i := 0; i < numberOfPartitions; i++ {
|
|
|
|
start := size * i
|
|
|
|
end := size * (i + 1)
|
|
|
|
if end > len(datums) {
|
|
|
|
end = len(datums)
|
|
|
|
}
|
|
|
|
|
|
|
|
partitions[i] = datums[start:end]
|
|
|
|
}
|
|
|
|
|
|
|
|
return partitions
|
|
|
|
}
|
|
|
|
|
2018-08-01 22:43:17 +00:00
|
|
|
// Make a MetricDatum from telegraf.Metric. It would check if all required fields of
|
|
|
|
// cloudwatch.StatisticSet are available. If so, it would build MetricDatum from statistic values.
|
|
|
|
// Otherwise, fields would still been built independently.
|
|
|
|
func BuildMetricDatum(buildStatistic bool, point telegraf.Metric) []*cloudwatch.MetricDatum {
|
2016-01-18 19:39:14 +00:00
|
|
|
|
2018-08-01 22:43:17 +00:00
|
|
|
fields := make(map[string]cloudwatchField)
|
2016-01-18 19:39:14 +00:00
|
|
|
|
|
|
|
for k, v := range point.Fields() {
|
|
|
|
|
2018-08-01 22:43:17 +00:00
|
|
|
val, ok := convert(v)
|
|
|
|
if !ok {
|
|
|
|
// Only fields with values that can be converted to float64 (and within CloudWatch boundary) are supported.
|
|
|
|
// Non-supported fields are skipped.
|
2017-08-28 23:56:03 +00:00
|
|
|
continue
|
|
|
|
}
|
2018-08-01 22:43:17 +00:00
|
|
|
|
|
|
|
sType, fieldName := getStatisticType(k)
|
|
|
|
|
|
|
|
// If statistic metric is not enabled or non-statistic type, just take current field as a value field.
|
|
|
|
if !buildStatistic || sType == statisticTypeNone {
|
|
|
|
fields[k] = &valueField{
|
|
|
|
metricName: point.Name(),
|
|
|
|
fieldName: k,
|
|
|
|
tags: point.Tags(),
|
|
|
|
timestamp: point.Time(),
|
|
|
|
value: val,
|
|
|
|
}
|
2017-08-28 23:56:03 +00:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2018-08-01 22:43:17 +00:00
|
|
|
// Otherwise, it shall be a statistic field.
|
|
|
|
if _, ok := fields[fieldName]; !ok {
|
|
|
|
// Hit an uncached field, create statisticField for first time
|
|
|
|
fields[fieldName] = &statisticField{
|
|
|
|
metricName: point.Name(),
|
|
|
|
fieldName: fieldName,
|
|
|
|
tags: point.Tags(),
|
|
|
|
timestamp: point.Time(),
|
|
|
|
values: map[statisticType]float64{
|
|
|
|
sType: val,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Add new statistic value to this field
|
|
|
|
fields[fieldName].addValue(sType, val)
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
2018-08-01 22:43:17 +00:00
|
|
|
}
|
2016-01-18 19:39:14 +00:00
|
|
|
|
2018-08-01 22:43:17 +00:00
|
|
|
var datums []*cloudwatch.MetricDatum
|
|
|
|
for _, f := range fields {
|
|
|
|
d := f.buildDatum()
|
|
|
|
datums = append(datums, d...)
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
return datums
|
|
|
|
}
|
|
|
|
|
|
|
|
// Make a list of Dimensions by using a Point's tags. CloudWatch supports up to
|
|
|
|
// 10 dimensions per metric so we only keep up to the first 10 alphabetically.
|
|
|
|
// This always includes the "host" tag if it exists.
|
2016-01-27 23:15:14 +00:00
|
|
|
func BuildDimensions(mTags map[string]string) []*cloudwatch.Dimension {
|
2016-01-18 19:39:14 +00:00
|
|
|
|
|
|
|
const MaxDimensions = 10
|
2016-01-27 23:15:14 +00:00
|
|
|
dimensions := make([]*cloudwatch.Dimension, int(math.Min(float64(len(mTags)), MaxDimensions)))
|
2016-01-18 19:39:14 +00:00
|
|
|
|
|
|
|
i := 0
|
|
|
|
|
|
|
|
// This is pretty ugly but we always want to include the "host" tag if it exists.
|
2016-01-27 23:15:14 +00:00
|
|
|
if host, ok := mTags["host"]; ok {
|
2016-01-18 19:39:14 +00:00
|
|
|
dimensions[i] = &cloudwatch.Dimension{
|
|
|
|
Name: aws.String("host"),
|
|
|
|
Value: aws.String(host),
|
|
|
|
}
|
|
|
|
i += 1
|
|
|
|
}
|
|
|
|
|
|
|
|
var keys []string
|
2016-01-27 23:15:14 +00:00
|
|
|
for k := range mTags {
|
2016-01-18 19:39:14 +00:00
|
|
|
if k != "host" {
|
|
|
|
keys = append(keys, k)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
sort.Strings(keys)
|
|
|
|
|
|
|
|
for _, k := range keys {
|
|
|
|
if i >= MaxDimensions {
|
|
|
|
break
|
|
|
|
}
|
|
|
|
|
|
|
|
dimensions[i] = &cloudwatch.Dimension{
|
|
|
|
Name: aws.String(k),
|
2016-01-27 23:15:14 +00:00
|
|
|
Value: aws.String(mTags[k]),
|
2016-01-18 19:39:14 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
i += 1
|
|
|
|
}
|
|
|
|
|
|
|
|
return dimensions
|
|
|
|
}
|
|
|
|
|
2018-08-01 22:43:17 +00:00
|
|
|
func getStatisticType(name string) (sType statisticType, fieldName string) {
|
|
|
|
switch {
|
|
|
|
case strings.HasSuffix(name, "_max"):
|
|
|
|
sType = statisticTypeMax
|
|
|
|
fieldName = strings.TrimSuffix(name, "_max")
|
|
|
|
case strings.HasSuffix(name, "_min"):
|
|
|
|
sType = statisticTypeMin
|
|
|
|
fieldName = strings.TrimSuffix(name, "_min")
|
|
|
|
case strings.HasSuffix(name, "_sum"):
|
|
|
|
sType = statisticTypeSum
|
|
|
|
fieldName = strings.TrimSuffix(name, "_sum")
|
|
|
|
case strings.HasSuffix(name, "_count"):
|
|
|
|
sType = statisticTypeCount
|
|
|
|
fieldName = strings.TrimSuffix(name, "_count")
|
|
|
|
default:
|
|
|
|
sType = statisticTypeNone
|
|
|
|
fieldName = name
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
func convert(v interface{}) (value float64, ok bool) {
|
|
|
|
|
|
|
|
ok = true
|
|
|
|
|
|
|
|
switch t := v.(type) {
|
|
|
|
case int:
|
|
|
|
value = float64(t)
|
|
|
|
case int32:
|
|
|
|
value = float64(t)
|
|
|
|
case int64:
|
|
|
|
value = float64(t)
|
|
|
|
case uint64:
|
|
|
|
value = float64(t)
|
|
|
|
case float64:
|
|
|
|
value = t
|
|
|
|
case bool:
|
|
|
|
if t {
|
|
|
|
value = 1
|
|
|
|
} else {
|
|
|
|
value = 0
|
|
|
|
}
|
|
|
|
case time.Time:
|
|
|
|
value = float64(t.Unix())
|
|
|
|
default:
|
|
|
|
// Skip unsupported type.
|
|
|
|
ok = false
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
// Do CloudWatch boundary checking
|
|
|
|
// Constraints at: http://docs.aws.amazon.com/AmazonCloudWatch/latest/APIReference/API_MetricDatum.html
|
|
|
|
switch {
|
|
|
|
case math.IsNaN(value):
|
|
|
|
return 0, false
|
|
|
|
case math.IsInf(value, 0):
|
|
|
|
return 0, false
|
|
|
|
case value > 0 && value < float64(8.515920e-109):
|
|
|
|
return 0, false
|
|
|
|
case value > float64(1.174271e+108):
|
|
|
|
return 0, false
|
|
|
|
}
|
|
|
|
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
2016-01-18 19:39:14 +00:00
|
|
|
func init() {
|
2016-01-27 21:21:36 +00:00
|
|
|
outputs.Add("cloudwatch", func() telegraf.Output {
|
2016-01-18 19:39:14 +00:00
|
|
|
return &CloudWatch{}
|
|
|
|
})
|
|
|
|
}
|