2017-08-10 19:36:11 +00:00
|
|
|
// +build !solaris
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
package tail
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
2018-09-18 16:23:45 +00:00
|
|
|
"log"
|
2017-06-16 20:16:48 +00:00
|
|
|
"strings"
|
2016-04-22 21:47:26 +00:00
|
|
|
"sync"
|
|
|
|
|
2017-03-29 21:25:33 +00:00
|
|
|
"github.com/influxdata/tail"
|
2019-07-12 00:39:59 +00:00
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
"github.com/influxdata/telegraf"
|
|
|
|
"github.com/influxdata/telegraf/internal/globpath"
|
|
|
|
"github.com/influxdata/telegraf/plugins/inputs"
|
|
|
|
"github.com/influxdata/telegraf/plugins/parsers"
|
|
|
|
)
|
|
|
|
|
2017-09-11 18:56:04 +00:00
|
|
|
const (
|
|
|
|
defaultWatchMethod = "inotify"
|
|
|
|
)
|
|
|
|
|
2019-07-12 00:39:59 +00:00
|
|
|
var (
|
|
|
|
offsets = make(map[string]int64)
|
|
|
|
offsetsMutex = new(sync.Mutex)
|
|
|
|
)
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
type Tail struct {
|
|
|
|
Files []string
|
|
|
|
FromBeginning bool
|
2016-12-16 14:01:49 +00:00
|
|
|
Pipe bool
|
2017-09-11 18:56:04 +00:00
|
|
|
WatchMethod string
|
2016-04-22 21:47:26 +00:00
|
|
|
|
2018-09-18 16:23:45 +00:00
|
|
|
tailers map[string]*tail.Tail
|
2019-07-12 00:39:59 +00:00
|
|
|
offsets map[string]int64
|
2018-09-18 16:23:45 +00:00
|
|
|
parserFunc parsers.ParserFunc
|
|
|
|
wg sync.WaitGroup
|
|
|
|
acc telegraf.Accumulator
|
2016-04-22 21:47:26 +00:00
|
|
|
|
|
|
|
sync.Mutex
|
|
|
|
}
|
|
|
|
|
2016-04-26 16:43:41 +00:00
|
|
|
func NewTail() *Tail {
|
2019-07-12 00:39:59 +00:00
|
|
|
offsetsMutex.Lock()
|
|
|
|
offsetsCopy := make(map[string]int64, len(offsets))
|
|
|
|
for k, v := range offsets {
|
|
|
|
offsetsCopy[k] = v
|
|
|
|
}
|
|
|
|
offsetsMutex.Unlock()
|
|
|
|
|
2016-04-26 16:43:41 +00:00
|
|
|
return &Tail{
|
|
|
|
FromBeginning: false,
|
2019-07-12 00:39:59 +00:00
|
|
|
offsets: offsetsCopy,
|
2016-04-26 16:43:41 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
const sampleConfig = `
|
|
|
|
## files to tail.
|
|
|
|
## These accept standard unix glob matching rules, but with the addition of
|
|
|
|
## ** as a "super asterisk". ie:
|
|
|
|
## "/var/log/**.log" -> recursively find all .log files in /var/log
|
|
|
|
## "/var/log/*/*.log" -> find all .log files with a parent dir in /var/log
|
|
|
|
## "/var/log/apache.log" -> just tail the apache log file
|
|
|
|
##
|
|
|
|
## See https://github.com/gobwas/glob for more examples
|
|
|
|
##
|
|
|
|
files = ["/var/mymetrics.out"]
|
|
|
|
## Read file from beginning.
|
|
|
|
from_beginning = false
|
2016-12-16 14:01:49 +00:00
|
|
|
## Whether file is a named pipe
|
|
|
|
pipe = false
|
2016-04-22 21:47:26 +00:00
|
|
|
|
2017-09-11 18:56:04 +00:00
|
|
|
## Method used to watch for file updates. Can be either "inotify" or "poll".
|
|
|
|
# watch_method = "inotify"
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
## Data format to consume.
|
2017-04-27 21:59:18 +00:00
|
|
|
## Each data format has its own unique set of configuration options, read
|
2016-04-22 21:47:26 +00:00
|
|
|
## more about them here:
|
|
|
|
## https://github.com/influxdata/telegraf/blob/master/docs/DATA_FORMATS_INPUT.md
|
|
|
|
data_format = "influx"
|
|
|
|
`
|
|
|
|
|
|
|
|
func (t *Tail) SampleConfig() string {
|
|
|
|
return sampleConfig
|
|
|
|
}
|
|
|
|
|
|
|
|
func (t *Tail) Description() string {
|
|
|
|
return "Stream a log file, like the tail -f command"
|
|
|
|
}
|
|
|
|
|
|
|
|
func (t *Tail) Gather(acc telegraf.Accumulator) error {
|
2018-09-17 23:20:50 +00:00
|
|
|
t.Lock()
|
|
|
|
defer t.Unlock()
|
|
|
|
|
|
|
|
return t.tailNewFiles(true)
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (t *Tail) Start(acc telegraf.Accumulator) error {
|
|
|
|
t.Lock()
|
|
|
|
defer t.Unlock()
|
|
|
|
|
|
|
|
t.acc = acc
|
2018-09-17 23:20:50 +00:00
|
|
|
t.tailers = make(map[string]*tail.Tail)
|
|
|
|
|
2019-07-12 00:39:59 +00:00
|
|
|
err := t.tailNewFiles(t.FromBeginning)
|
|
|
|
|
|
|
|
// clear offsets
|
|
|
|
t.offsets = make(map[string]int64)
|
|
|
|
// assumption that once Start is called, all parallel plugins have already been initialized
|
|
|
|
offsetsMutex.Lock()
|
|
|
|
offsets = make(map[string]int64)
|
|
|
|
offsetsMutex.Unlock()
|
|
|
|
|
|
|
|
return err
|
2018-09-17 23:20:50 +00:00
|
|
|
}
|
2016-04-22 21:47:26 +00:00
|
|
|
|
2018-09-17 23:20:50 +00:00
|
|
|
func (t *Tail) tailNewFiles(fromBeginning bool) error {
|
2017-09-11 18:56:04 +00:00
|
|
|
var poll bool
|
|
|
|
if t.WatchMethod == "poll" {
|
|
|
|
poll = true
|
|
|
|
}
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
// Create a "tailer" for each file
|
|
|
|
for _, filepath := range t.Files {
|
|
|
|
g, err := globpath.Compile(filepath)
|
|
|
|
if err != nil {
|
2019-07-12 00:39:59 +00:00
|
|
|
t.acc.AddError(fmt.Errorf("glob %s failed to compile, %s", filepath, err))
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
2018-12-18 22:23:25 +00:00
|
|
|
for _, file := range g.Match() {
|
2018-09-17 23:20:50 +00:00
|
|
|
if _, ok := t.tailers[file]; ok {
|
|
|
|
// we're already tailing this file
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2019-07-12 00:39:59 +00:00
|
|
|
var seek *tail.SeekInfo
|
|
|
|
if !t.Pipe && !fromBeginning {
|
|
|
|
if offset, ok := t.offsets[file]; ok {
|
|
|
|
log.Printf("D! [inputs.tail] using offset %d for file: %v", offset, file)
|
|
|
|
seek = &tail.SeekInfo{
|
|
|
|
Whence: 0,
|
|
|
|
Offset: offset,
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
seek = &tail.SeekInfo{
|
|
|
|
Whence: 2,
|
|
|
|
Offset: 0,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
tailer, err := tail.TailFile(file,
|
|
|
|
tail.Config{
|
2016-07-18 13:44:25 +00:00
|
|
|
ReOpen: true,
|
|
|
|
Follow: true,
|
2016-12-16 14:01:49 +00:00
|
|
|
Location: seek,
|
2016-07-18 13:44:25 +00:00
|
|
|
MustExist: true,
|
2017-09-11 18:56:04 +00:00
|
|
|
Poll: poll,
|
2016-12-16 14:01:49 +00:00
|
|
|
Pipe: t.Pipe,
|
2017-08-16 19:06:07 +00:00
|
|
|
Logger: tail.DiscardingLogger,
|
2016-04-22 21:47:26 +00:00
|
|
|
})
|
|
|
|
if err != nil {
|
2018-09-17 23:20:50 +00:00
|
|
|
t.acc.AddError(err)
|
2016-04-22 21:47:26 +00:00
|
|
|
continue
|
|
|
|
}
|
2018-09-18 16:23:45 +00:00
|
|
|
|
|
|
|
log.Printf("D! [inputs.tail] tail added for file: %v", file)
|
|
|
|
|
|
|
|
parser, err := t.parserFunc()
|
|
|
|
if err != nil {
|
|
|
|
t.acc.AddError(fmt.Errorf("error creating parser: %v", err))
|
|
|
|
}
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
// create a goroutine for each "tailer"
|
2016-05-24 14:16:33 +00:00
|
|
|
t.wg.Add(1)
|
2018-09-18 16:23:45 +00:00
|
|
|
go t.receiver(parser, tailer)
|
|
|
|
t.tailers[tailer.Filename] = tailer
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// this is launched as a goroutine to continuously watch a tailed logfile
|
|
|
|
// for changes, parse any incoming msgs, and add to the accumulator.
|
2018-09-18 16:23:45 +00:00
|
|
|
func (t *Tail) receiver(parser parsers.Parser, tailer *tail.Tail) {
|
2016-04-22 21:47:26 +00:00
|
|
|
defer t.wg.Done()
|
|
|
|
|
2018-09-18 16:23:45 +00:00
|
|
|
var firstLine = true
|
|
|
|
var metrics []telegraf.Metric
|
2016-04-22 21:47:26 +00:00
|
|
|
var m telegraf.Metric
|
|
|
|
var err error
|
|
|
|
var line *tail.Line
|
|
|
|
for line = range tailer.Lines {
|
|
|
|
if line.Err != nil {
|
2019-07-12 00:39:59 +00:00
|
|
|
t.acc.AddError(fmt.Errorf("error tailing file %s, Error: %s", tailer.Filename, err))
|
2016-04-22 21:47:26 +00:00
|
|
|
continue
|
|
|
|
}
|
2017-06-16 20:16:48 +00:00
|
|
|
// Fix up files with Windows line endings.
|
|
|
|
text := strings.TrimRight(line.Text, "\r")
|
|
|
|
|
2018-09-18 16:23:45 +00:00
|
|
|
if firstLine {
|
|
|
|
metrics, err = parser.Parse([]byte(text))
|
|
|
|
if err == nil {
|
|
|
|
if len(metrics) == 0 {
|
|
|
|
firstLine = false
|
|
|
|
continue
|
|
|
|
} else {
|
|
|
|
m = metrics[0]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
firstLine = false
|
|
|
|
} else {
|
|
|
|
m, err = parser.ParseLine(text)
|
|
|
|
}
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
if err == nil {
|
2018-06-22 00:55:54 +00:00
|
|
|
if m != nil {
|
|
|
|
tags := m.Tags()
|
|
|
|
tags["path"] = tailer.Filename
|
|
|
|
t.acc.AddFields(m.Name(), m.Fields(), tags, m.Time())
|
|
|
|
}
|
2016-04-22 21:47:26 +00:00
|
|
|
} else {
|
2019-07-12 00:39:59 +00:00
|
|
|
t.acc.AddError(fmt.Errorf("malformed log line in %s: [%s], Error: %s",
|
2017-03-24 19:03:36 +00:00
|
|
|
tailer.Filename, line.Text, err))
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
|
|
|
}
|
2018-09-18 16:23:45 +00:00
|
|
|
|
|
|
|
log.Printf("D! [inputs.tail] tail removed for file: %v", tailer.Filename)
|
|
|
|
|
2016-12-16 14:01:49 +00:00
|
|
|
if err := tailer.Err(); err != nil {
|
2019-07-12 00:39:59 +00:00
|
|
|
t.acc.AddError(fmt.Errorf("error tailing file %s, Error: %s", tailer.Filename, err))
|
2016-12-16 14:01:49 +00:00
|
|
|
}
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (t *Tail) Stop() {
|
|
|
|
t.Lock()
|
|
|
|
defer t.Unlock()
|
|
|
|
|
2017-03-24 19:03:36 +00:00
|
|
|
for _, tailer := range t.tailers {
|
2019-07-12 00:39:59 +00:00
|
|
|
if !t.Pipe && !t.FromBeginning {
|
|
|
|
// store offset for resume
|
|
|
|
offset, err := tailer.Tell()
|
|
|
|
if err == nil {
|
|
|
|
log.Printf("D! [inputs.tail] recording offset %d for file: %v", offset, tailer.Filename)
|
|
|
|
} else {
|
|
|
|
t.acc.AddError(fmt.Errorf("error recording offset for file %s", tailer.Filename))
|
|
|
|
}
|
|
|
|
}
|
2017-03-24 19:03:36 +00:00
|
|
|
err := tailer.Stop()
|
2016-04-22 21:47:26 +00:00
|
|
|
if err != nil {
|
2019-07-12 00:39:59 +00:00
|
|
|
t.acc.AddError(fmt.Errorf("error stopping tail on file %s", tailer.Filename))
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
2018-09-18 16:23:45 +00:00
|
|
|
}
|
|
|
|
|
2016-04-22 21:47:26 +00:00
|
|
|
t.wg.Wait()
|
2019-07-12 00:39:59 +00:00
|
|
|
|
|
|
|
// persist offsets
|
|
|
|
offsetsMutex.Lock()
|
|
|
|
for k, v := range t.offsets {
|
|
|
|
offsets[k] = v
|
|
|
|
}
|
|
|
|
offsetsMutex.Unlock()
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
|
|
|
|
2018-09-18 16:23:45 +00:00
|
|
|
func (t *Tail) SetParserFunc(fn parsers.ParserFunc) {
|
|
|
|
t.parserFunc = fn
|
2016-04-22 21:47:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
inputs.Add("tail", func() telegraf.Input {
|
2016-04-26 16:43:41 +00:00
|
|
|
return NewTail()
|
2016-04-22 21:47:26 +00:00
|
|
|
})
|
|
|
|
}
|