2016-01-20 23:21:19 +00:00
|
|
|
package system
|
|
|
|
|
|
|
|
import (
|
2016-02-24 04:58:14 +00:00
|
|
|
"encoding/json"
|
2016-01-20 23:21:19 +00:00
|
|
|
"fmt"
|
2016-01-29 17:11:28 +00:00
|
|
|
"log"
|
2016-02-24 04:58:14 +00:00
|
|
|
"regexp"
|
|
|
|
"strconv"
|
2016-01-20 23:21:19 +00:00
|
|
|
"strings"
|
|
|
|
"sync"
|
|
|
|
"time"
|
|
|
|
|
2016-01-27 21:21:36 +00:00
|
|
|
"github.com/influxdata/telegraf"
|
2016-01-20 23:21:19 +00:00
|
|
|
"github.com/influxdata/telegraf/plugins/inputs"
|
|
|
|
|
|
|
|
"github.com/fsouza/go-dockerclient"
|
2016-03-22 17:21:58 +00:00
|
|
|
"errors"
|
2016-01-20 23:21:19 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
type Docker struct {
|
|
|
|
Endpoint string
|
|
|
|
ContainerNames []string
|
|
|
|
|
2016-03-22 17:21:58 +00:00
|
|
|
// Enables TLS
|
|
|
|
TLSEnabled bool `toml:"tls_enabled"`
|
|
|
|
// Path to CA file
|
|
|
|
TLSCA string `toml:"tls_ca"`
|
|
|
|
// Path to cert file
|
|
|
|
TLSCert string `toml:"tls_cert"`
|
|
|
|
// Path to cert key file
|
|
|
|
TLSKey string `toml:"tls_key"`
|
|
|
|
|
2016-02-24 04:58:14 +00:00
|
|
|
client DockerClient
|
2016-01-20 23:21:19 +00:00
|
|
|
}
|
|
|
|
|
2016-02-24 04:58:14 +00:00
|
|
|
type DockerClient interface {
|
|
|
|
// Docker Client wrapper
|
|
|
|
// Useful for test
|
|
|
|
Info() (*docker.Env, error)
|
|
|
|
ListContainers(opts docker.ListContainersOptions) ([]docker.APIContainers, error)
|
|
|
|
Stats(opts docker.StatsOptions) error
|
|
|
|
}
|
|
|
|
|
|
|
|
const (
|
|
|
|
KB = 1000
|
|
|
|
MB = 1000 * KB
|
|
|
|
GB = 1000 * MB
|
|
|
|
TB = 1000 * GB
|
|
|
|
PB = 1000 * TB
|
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
sizeRegex = regexp.MustCompile(`^(\d+(\.\d+)*) ?([kKmMgGtTpP])?[bB]?$`)
|
|
|
|
)
|
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
var sampleConfig = `
|
2016-02-18 21:26:51 +00:00
|
|
|
## Docker Endpoint
|
|
|
|
## To use TCP, set endpoint = "tcp://[ip]:[port]"
|
|
|
|
## To use environment variables (ie, docker-machine), set endpoint = "ENV"
|
2016-01-20 23:21:19 +00:00
|
|
|
endpoint = "unix:///var/run/docker.sock"
|
2016-03-22 17:21:58 +00:00
|
|
|
## To collect metrics from TLS-enabled daemon
|
|
|
|
# tls_enabled = true
|
|
|
|
# tls_ca = "~/certificates_path/ca.pem"
|
|
|
|
# tls_cert = "~/certificates_path/cert.pem"
|
|
|
|
# tls_key = "~/certificates_path/key.pem"
|
2016-02-18 21:26:51 +00:00
|
|
|
## Only collect metrics for these containers, collect all if empty
|
2016-01-20 23:21:19 +00:00
|
|
|
container_names = []
|
|
|
|
`
|
|
|
|
|
|
|
|
func (d *Docker) Description() string {
|
|
|
|
return "Read metrics about docker containers"
|
|
|
|
}
|
|
|
|
|
|
|
|
func (d *Docker) SampleConfig() string { return sampleConfig }
|
|
|
|
|
2016-01-27 21:21:36 +00:00
|
|
|
func (d *Docker) Gather(acc telegraf.Accumulator) error {
|
2016-01-20 23:21:19 +00:00
|
|
|
if d.client == nil {
|
|
|
|
var c *docker.Client
|
|
|
|
var err error
|
|
|
|
if d.Endpoint == "ENV" {
|
|
|
|
c, err = docker.NewClientFromEnv()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
} else if d.Endpoint == "" {
|
|
|
|
c, err = docker.NewClient("unix:///var/run/docker.sock")
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
} else {
|
2016-03-22 17:21:58 +00:00
|
|
|
if !d.TLSEnabled {
|
|
|
|
c, err = docker.NewClient(d.Endpoint)
|
|
|
|
} else {
|
|
|
|
if d.TLSCert == "" {
|
|
|
|
return errors.New("tls_cert must be configured when tls_enable is set to true");
|
|
|
|
}
|
|
|
|
|
|
|
|
if d.TLSKey == "" {
|
|
|
|
return errors.New("tls_key must be configured when tls_enable is set to true");
|
|
|
|
}
|
|
|
|
|
|
|
|
if d.TLSCA == "" {
|
|
|
|
return errors.New("tls_ca must be configured when tls_enable is set to true");
|
|
|
|
}
|
|
|
|
|
|
|
|
c, err = docker.NewTLSClient(d.Endpoint, d.TLSCert, d.TLSKey, d.TLSCA)
|
|
|
|
}
|
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
d.client = c
|
|
|
|
}
|
|
|
|
|
2016-02-24 04:58:14 +00:00
|
|
|
// Get daemon info
|
|
|
|
err := d.gatherInfo(acc)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Println(err.Error())
|
|
|
|
}
|
|
|
|
|
|
|
|
// List containers
|
2016-01-20 23:21:19 +00:00
|
|
|
opts := docker.ListContainersOptions{}
|
|
|
|
containers, err := d.client.ListContainers(opts)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2016-02-24 04:58:14 +00:00
|
|
|
// Get container data
|
2016-01-20 23:21:19 +00:00
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(len(containers))
|
|
|
|
for _, container := range containers {
|
2016-02-09 15:20:56 +00:00
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
go func(c docker.APIContainers) {
|
|
|
|
defer wg.Done()
|
|
|
|
err := d.gatherContainer(c, acc)
|
|
|
|
if err != nil {
|
|
|
|
fmt.Println(err.Error())
|
|
|
|
}
|
|
|
|
}(container)
|
|
|
|
}
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2016-02-24 04:58:14 +00:00
|
|
|
func (d *Docker) gatherInfo(acc telegraf.Accumulator) error {
|
|
|
|
// Init vars
|
|
|
|
var driverStatus [][]string
|
|
|
|
dataFields := make(map[string]interface{})
|
|
|
|
metadataFields := make(map[string]interface{})
|
|
|
|
now := time.Now()
|
|
|
|
// Get info from docker daemon
|
|
|
|
info, err := d.client.Info()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
fields := map[string]interface{}{
|
|
|
|
"n_cpus": info.GetInt64("NCPU"),
|
|
|
|
"n_used_file_descriptors": info.GetInt64("NFd"),
|
|
|
|
"n_containers": info.GetInt64("Containers"),
|
|
|
|
"n_images": info.GetInt64("Images"),
|
|
|
|
"n_goroutines": info.GetInt64("NGoroutines"),
|
|
|
|
"n_listener_events": info.GetInt64("NEventsListener"),
|
|
|
|
}
|
|
|
|
// Add metrics
|
|
|
|
acc.AddFields("docker",
|
|
|
|
fields,
|
|
|
|
nil,
|
|
|
|
now)
|
|
|
|
acc.AddFields("docker",
|
|
|
|
map[string]interface{}{"memory_total": info.GetInt64("MemTotal")},
|
|
|
|
map[string]string{"unit": "bytes"},
|
|
|
|
now)
|
|
|
|
// Get storage metrics
|
|
|
|
driverStatusRaw := []byte(info.Get("DriverStatus"))
|
|
|
|
json.Unmarshal(driverStatusRaw, &driverStatus)
|
|
|
|
for _, rawData := range driverStatus {
|
|
|
|
// Try to convert string to int (bytes)
|
|
|
|
value, err := parseSize(rawData[1])
|
|
|
|
if err != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
name := strings.ToLower(strings.Replace(rawData[0], " ", "_", -1))
|
|
|
|
if name == "pool_blocksize" {
|
|
|
|
// pool blocksize
|
|
|
|
acc.AddFields("docker",
|
|
|
|
map[string]interface{}{"pool_blocksize": value},
|
|
|
|
map[string]string{"unit": "bytes"},
|
|
|
|
now)
|
|
|
|
} else if strings.HasPrefix(name, "data_space_") {
|
|
|
|
// data space
|
|
|
|
field_name := strings.TrimPrefix(name, "data_space_")
|
|
|
|
dataFields[field_name] = value
|
|
|
|
} else if strings.HasPrefix(name, "metadata_space_") {
|
|
|
|
// metadata space
|
|
|
|
field_name := strings.TrimPrefix(name, "metadata_space_")
|
|
|
|
metadataFields[field_name] = value
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(dataFields) > 0 {
|
|
|
|
acc.AddFields("docker_data",
|
|
|
|
dataFields,
|
|
|
|
map[string]string{"unit": "bytes"},
|
|
|
|
now)
|
|
|
|
}
|
|
|
|
if len(metadataFields) > 0 {
|
|
|
|
acc.AddFields("docker_metadata",
|
|
|
|
metadataFields,
|
|
|
|
map[string]string{"unit": "bytes"},
|
|
|
|
now)
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
func (d *Docker) gatherContainer(
|
|
|
|
container docker.APIContainers,
|
2016-01-27 21:21:36 +00:00
|
|
|
acc telegraf.Accumulator,
|
2016-01-20 23:21:19 +00:00
|
|
|
) error {
|
|
|
|
// Parse container name
|
|
|
|
cname := "unknown"
|
|
|
|
if len(container.Names) > 0 {
|
|
|
|
// Not sure what to do with other names, just take the first.
|
|
|
|
cname = strings.TrimPrefix(container.Names[0], "/")
|
|
|
|
}
|
|
|
|
|
|
|
|
tags := map[string]string{
|
|
|
|
"cont_id": container.ID,
|
|
|
|
"cont_name": cname,
|
|
|
|
"cont_image": container.Image,
|
|
|
|
}
|
|
|
|
if len(d.ContainerNames) > 0 {
|
|
|
|
if !sliceContains(cname, d.ContainerNames) {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
statChan := make(chan *docker.Stats)
|
|
|
|
done := make(chan bool)
|
|
|
|
statOpts := docker.StatsOptions{
|
|
|
|
Stream: false,
|
|
|
|
ID: container.ID,
|
|
|
|
Stats: statChan,
|
|
|
|
Done: done,
|
|
|
|
Timeout: time.Duration(time.Second * 5),
|
|
|
|
}
|
|
|
|
|
|
|
|
go func() {
|
2016-01-29 17:11:28 +00:00
|
|
|
err := d.client.Stats(statOpts)
|
|
|
|
if err != nil {
|
|
|
|
log.Printf("Error getting docker stats: %s\n", err.Error())
|
|
|
|
}
|
2016-01-20 23:21:19 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
stat := <-statChan
|
2016-01-22 18:54:12 +00:00
|
|
|
close(done)
|
2016-01-20 23:21:19 +00:00
|
|
|
|
2016-01-29 17:11:28 +00:00
|
|
|
if stat == nil {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
// Add labels to tags
|
|
|
|
for k, v := range container.Labels {
|
|
|
|
tags[k] = v
|
|
|
|
}
|
|
|
|
|
|
|
|
gatherContainerStats(stat, acc, tags)
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func gatherContainerStats(
|
|
|
|
stat *docker.Stats,
|
2016-01-27 21:21:36 +00:00
|
|
|
acc telegraf.Accumulator,
|
2016-01-20 23:21:19 +00:00
|
|
|
tags map[string]string,
|
|
|
|
) {
|
|
|
|
now := stat.Read
|
|
|
|
|
|
|
|
memfields := map[string]interface{}{
|
|
|
|
"max_usage": stat.MemoryStats.MaxUsage,
|
|
|
|
"usage": stat.MemoryStats.Usage,
|
|
|
|
"fail_count": stat.MemoryStats.Failcnt,
|
|
|
|
"limit": stat.MemoryStats.Limit,
|
|
|
|
"total_pgmafault": stat.MemoryStats.Stats.TotalPgmafault,
|
|
|
|
"cache": stat.MemoryStats.Stats.Cache,
|
|
|
|
"mapped_file": stat.MemoryStats.Stats.MappedFile,
|
|
|
|
"total_inactive_file": stat.MemoryStats.Stats.TotalInactiveFile,
|
|
|
|
"pgpgout": stat.MemoryStats.Stats.Pgpgout,
|
|
|
|
"rss": stat.MemoryStats.Stats.Rss,
|
|
|
|
"total_mapped_file": stat.MemoryStats.Stats.TotalMappedFile,
|
|
|
|
"writeback": stat.MemoryStats.Stats.Writeback,
|
|
|
|
"unevictable": stat.MemoryStats.Stats.Unevictable,
|
|
|
|
"pgpgin": stat.MemoryStats.Stats.Pgpgin,
|
|
|
|
"total_unevictable": stat.MemoryStats.Stats.TotalUnevictable,
|
|
|
|
"pgmajfault": stat.MemoryStats.Stats.Pgmajfault,
|
|
|
|
"total_rss": stat.MemoryStats.Stats.TotalRss,
|
|
|
|
"total_rss_huge": stat.MemoryStats.Stats.TotalRssHuge,
|
|
|
|
"total_writeback": stat.MemoryStats.Stats.TotalWriteback,
|
|
|
|
"total_inactive_anon": stat.MemoryStats.Stats.TotalInactiveAnon,
|
|
|
|
"rss_huge": stat.MemoryStats.Stats.RssHuge,
|
|
|
|
"hierarchical_memory_limit": stat.MemoryStats.Stats.HierarchicalMemoryLimit,
|
|
|
|
"total_pgfault": stat.MemoryStats.Stats.TotalPgfault,
|
|
|
|
"total_active_file": stat.MemoryStats.Stats.TotalActiveFile,
|
|
|
|
"active_anon": stat.MemoryStats.Stats.ActiveAnon,
|
|
|
|
"total_active_anon": stat.MemoryStats.Stats.TotalActiveAnon,
|
|
|
|
"total_pgpgout": stat.MemoryStats.Stats.TotalPgpgout,
|
|
|
|
"total_cache": stat.MemoryStats.Stats.TotalCache,
|
|
|
|
"inactive_anon": stat.MemoryStats.Stats.InactiveAnon,
|
|
|
|
"active_file": stat.MemoryStats.Stats.ActiveFile,
|
|
|
|
"pgfault": stat.MemoryStats.Stats.Pgfault,
|
|
|
|
"inactive_file": stat.MemoryStats.Stats.InactiveFile,
|
|
|
|
"total_pgpgin": stat.MemoryStats.Stats.TotalPgpgin,
|
2016-02-11 10:49:48 +00:00
|
|
|
"usage_percent": calculateMemPercent(stat),
|
2016-01-20 23:21:19 +00:00
|
|
|
}
|
|
|
|
acc.AddFields("docker_mem", memfields, tags, now)
|
|
|
|
|
|
|
|
cpufields := map[string]interface{}{
|
|
|
|
"usage_total": stat.CPUStats.CPUUsage.TotalUsage,
|
|
|
|
"usage_in_usermode": stat.CPUStats.CPUUsage.UsageInUsermode,
|
|
|
|
"usage_in_kernelmode": stat.CPUStats.CPUUsage.UsageInKernelmode,
|
|
|
|
"usage_system": stat.CPUStats.SystemCPUUsage,
|
|
|
|
"throttling_periods": stat.CPUStats.ThrottlingData.Periods,
|
|
|
|
"throttling_throttled_periods": stat.CPUStats.ThrottlingData.ThrottledPeriods,
|
|
|
|
"throttling_throttled_time": stat.CPUStats.ThrottlingData.ThrottledTime,
|
2016-02-11 10:49:48 +00:00
|
|
|
"usage_percent": calculateCPUPercent(stat),
|
2016-01-20 23:21:19 +00:00
|
|
|
}
|
|
|
|
cputags := copyTags(tags)
|
|
|
|
cputags["cpu"] = "cpu-total"
|
|
|
|
acc.AddFields("docker_cpu", cpufields, cputags, now)
|
|
|
|
|
|
|
|
for i, percpu := range stat.CPUStats.CPUUsage.PercpuUsage {
|
|
|
|
percputags := copyTags(tags)
|
|
|
|
percputags["cpu"] = fmt.Sprintf("cpu%d", i)
|
|
|
|
acc.AddFields("docker_cpu", map[string]interface{}{"usage_total": percpu}, percputags, now)
|
|
|
|
}
|
|
|
|
|
|
|
|
for network, netstats := range stat.Networks {
|
|
|
|
netfields := map[string]interface{}{
|
|
|
|
"rx_dropped": netstats.RxDropped,
|
|
|
|
"rx_bytes": netstats.RxBytes,
|
|
|
|
"rx_errors": netstats.RxErrors,
|
|
|
|
"tx_packets": netstats.TxPackets,
|
|
|
|
"tx_dropped": netstats.TxDropped,
|
|
|
|
"rx_packets": netstats.RxPackets,
|
|
|
|
"tx_errors": netstats.TxErrors,
|
|
|
|
"tx_bytes": netstats.TxBytes,
|
|
|
|
}
|
|
|
|
// Create a new network tag dictionary for the "network" tag
|
|
|
|
nettags := copyTags(tags)
|
|
|
|
nettags["network"] = network
|
|
|
|
acc.AddFields("docker_net", netfields, nettags, now)
|
|
|
|
}
|
|
|
|
|
|
|
|
gatherBlockIOMetrics(stat, acc, tags, now)
|
|
|
|
}
|
|
|
|
|
2016-02-11 10:49:48 +00:00
|
|
|
func calculateMemPercent(stat *docker.Stats) float64 {
|
|
|
|
var memPercent = 0.0
|
|
|
|
if stat.MemoryStats.Limit > 0 {
|
|
|
|
memPercent = float64(stat.MemoryStats.Usage) / float64(stat.MemoryStats.Limit) * 100.0
|
|
|
|
}
|
|
|
|
return memPercent
|
|
|
|
}
|
|
|
|
|
2016-02-09 15:20:56 +00:00
|
|
|
func calculateCPUPercent(stat *docker.Stats) float64 {
|
2016-02-11 10:49:48 +00:00
|
|
|
var cpuPercent = 0.0
|
|
|
|
// calculate the change for the cpu and system usage of the container in between readings
|
|
|
|
cpuDelta := float64(stat.CPUStats.CPUUsage.TotalUsage) - float64(stat.PreCPUStats.CPUUsage.TotalUsage)
|
|
|
|
systemDelta := float64(stat.CPUStats.SystemCPUUsage) - float64(stat.PreCPUStats.SystemCPUUsage)
|
2016-02-09 15:20:56 +00:00
|
|
|
|
|
|
|
if systemDelta > 0.0 && cpuDelta > 0.0 {
|
|
|
|
cpuPercent = (cpuDelta / systemDelta) * float64(len(stat.CPUStats.CPUUsage.PercpuUsage)) * 100.0
|
|
|
|
}
|
|
|
|
return cpuPercent
|
|
|
|
}
|
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
func gatherBlockIOMetrics(
|
|
|
|
stat *docker.Stats,
|
2016-01-27 21:21:36 +00:00
|
|
|
acc telegraf.Accumulator,
|
2016-01-20 23:21:19 +00:00
|
|
|
tags map[string]string,
|
|
|
|
now time.Time,
|
|
|
|
) {
|
|
|
|
blkioStats := stat.BlkioStats
|
|
|
|
// Make a map of devices to their block io stats
|
|
|
|
deviceStatMap := make(map[string]map[string]interface{})
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOServiceBytesRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
_, ok := deviceStatMap[device]
|
|
|
|
if !ok {
|
|
|
|
deviceStatMap[device] = make(map[string]interface{})
|
|
|
|
}
|
|
|
|
|
|
|
|
field := fmt.Sprintf("io_service_bytes_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOServicedRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
_, ok := deviceStatMap[device]
|
|
|
|
if !ok {
|
|
|
|
deviceStatMap[device] = make(map[string]interface{})
|
|
|
|
}
|
|
|
|
|
|
|
|
field := fmt.Sprintf("io_serviced_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOQueueRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
field := fmt.Sprintf("io_queue_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOServiceTimeRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
field := fmt.Sprintf("io_service_time_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOWaitTimeRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
field := fmt.Sprintf("io_wait_time_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOMergedRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
field := fmt.Sprintf("io_merged_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.IOTimeRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
field := fmt.Sprintf("io_time_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for _, metric := range blkioStats.SectorsRecursive {
|
|
|
|
device := fmt.Sprintf("%d:%d", metric.Major, metric.Minor)
|
|
|
|
field := fmt.Sprintf("sectors_recursive_%s", strings.ToLower(metric.Op))
|
|
|
|
deviceStatMap[device][field] = metric.Value
|
|
|
|
}
|
|
|
|
|
|
|
|
for device, fields := range deviceStatMap {
|
|
|
|
iotags := copyTags(tags)
|
|
|
|
iotags["device"] = device
|
|
|
|
acc.AddFields("docker_blkio", fields, iotags, now)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func copyTags(in map[string]string) map[string]string {
|
|
|
|
out := make(map[string]string)
|
|
|
|
for k, v := range in {
|
|
|
|
out[k] = v
|
|
|
|
}
|
|
|
|
return out
|
|
|
|
}
|
|
|
|
|
|
|
|
func sliceContains(in string, sl []string) bool {
|
|
|
|
for _, str := range sl {
|
|
|
|
if str == in {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2016-02-24 04:58:14 +00:00
|
|
|
// Parses the human-readable size string into the amount it represents.
|
|
|
|
func parseSize(sizeStr string) (int64, error) {
|
|
|
|
matches := sizeRegex.FindStringSubmatch(sizeStr)
|
|
|
|
if len(matches) != 4 {
|
|
|
|
return -1, fmt.Errorf("invalid size: '%s'", sizeStr)
|
|
|
|
}
|
|
|
|
|
|
|
|
size, err := strconv.ParseFloat(matches[1], 64)
|
|
|
|
if err != nil {
|
|
|
|
return -1, err
|
|
|
|
}
|
|
|
|
|
|
|
|
uMap := map[string]int64{"k": KB, "m": MB, "g": GB, "t": TB, "p": PB}
|
|
|
|
unitPrefix := strings.ToLower(matches[3])
|
|
|
|
if mul, ok := uMap[unitPrefix]; ok {
|
|
|
|
size *= float64(mul)
|
|
|
|
}
|
|
|
|
|
|
|
|
return int64(size), nil
|
|
|
|
}
|
|
|
|
|
2016-01-20 23:21:19 +00:00
|
|
|
func init() {
|
2016-01-27 21:21:36 +00:00
|
|
|
inputs.Add("docker", func() telegraf.Input {
|
2016-01-20 23:21:19 +00:00
|
|
|
return &Docker{}
|
|
|
|
})
|
|
|
|
}
|