Add Solr input plugin (#2019)
This commit is contained in:
parent
293ca69a47
commit
89650bdf52
|
@ -195,6 +195,7 @@ configuration options.
|
|||
* [smart](./plugins/inputs/smart)
|
||||
* [snmp](./plugins/inputs/snmp)
|
||||
* [snmp_legacy](./plugins/inputs/snmp_legacy)
|
||||
* [solr](./plugins/inputs/solr)
|
||||
* [sql server](./plugins/inputs/sqlserver) (microsoft)
|
||||
* [teamspeak](./plugins/inputs/teamspeak)
|
||||
* [tomcat](./plugins/inputs/tomcat)
|
||||
|
|
|
@ -2329,6 +2329,15 @@
|
|||
# sub_tables=[".1.3.6.1.2.1.2.2.1.13", "bytes_recv", "bytes_send"]
|
||||
|
||||
|
||||
# # Read metrics from Solr Server
|
||||
# [[inputs.solr]]
|
||||
# ## specify a list of one or more Solr servers
|
||||
# servers = ["http://localhost:8983"]
|
||||
# ##
|
||||
# ## specify a list of one or more Solr cores (default - all)
|
||||
# # cores = ["main"]
|
||||
|
||||
|
||||
# # Read metrics from Microsoft SQL Server
|
||||
# [[inputs.sqlserver]]
|
||||
# ## Specify instances to monitor with a list of connection strings.
|
||||
|
|
|
@ -80,6 +80,7 @@ import (
|
|||
_ "github.com/influxdata/telegraf/plugins/inputs/snmp"
|
||||
_ "github.com/influxdata/telegraf/plugins/inputs/snmp_legacy"
|
||||
_ "github.com/influxdata/telegraf/plugins/inputs/socket_listener"
|
||||
_ "github.com/influxdata/telegraf/plugins/inputs/solr"
|
||||
_ "github.com/influxdata/telegraf/plugins/inputs/sqlserver"
|
||||
_ "github.com/influxdata/telegraf/plugins/inputs/statsd"
|
||||
_ "github.com/influxdata/telegraf/plugins/inputs/sysstat"
|
||||
|
|
|
@ -0,0 +1,31 @@
|
|||
# Solr input plugin
|
||||
|
||||
The [solr](http://lucene.apache.org/solr/) plugin collects stats via the
|
||||
[MBean Request Handler](https://cwiki.apache.org/confluence/display/solr/MBean+Request+Handler)
|
||||
|
||||
More about [performance statistics](https://cwiki.apache.org/confluence/display/solr/Performance+Statistics+Reference)
|
||||
|
||||
### Configuration:
|
||||
|
||||
```
|
||||
[[inputs.solr]]
|
||||
## specify a list of one or more Solr servers
|
||||
servers = ["http://localhost:8983"]
|
||||
##
|
||||
## specify a list of one or more Solr cores (default - all)
|
||||
# cores = ["main"]
|
||||
```
|
||||
|
||||
### Example output of gathered metrics:
|
||||
|
||||
```
|
||||
➜ ~ telegraf -config telegraf.conf -input-filter solr -test
|
||||
* Plugin: solr, Collection 1
|
||||
> solr_core,core=main,handler=searcher,host=testhost deleted_docs=17616645i,max_docs=261848363i,num_docs=244231718i 1478214949000000000
|
||||
> solr_core,core=main,handler=core,host=testhost deleted_docs=0i,max_docs=0i,num_docs=0i 1478214949000000000
|
||||
> solr_queryhandler,core=main,handler=/replication,host=testhost 15min_rate_reqs_per_second=0.000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000444659081257,5min_rate_reqs_per_second=0.00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000014821969375,75th_pc_request_time=16.484211,95th_pc_request_time=16.484211,999th_pc_request_time=16.484211,99th_pc_request_time=16.484211,avg_requests_per_second=0.0000008443809966322143,avg_time_per_request=12.984811,errors=0i,handler_start=1474662050865i,median_request_time=11.352427,requests=3i,timeouts=0i,total_time=38.954433 1478214949000000000
|
||||
> solr_queryhandler,core=main,handler=/update/extract,host=testhost 15min_rate_reqs_per_second=0,5min_rate_reqs_per_second=0,75th_pc_request_time=0,95th_pc_request_time=0,999th_pc_request_time=0,99th_pc_request_time=0,avg_requests_per_second=0,avg_time_per_request=0,errors=0i,handler_start=0i,median_request_time=0,requests=0i,timeouts=0i,total_time=0 1478214949000000000
|
||||
> solr_queryhandler,core=main,handler=org.apache.solr.handler.component.SearchHandler,host=testhost 15min_rate_reqs_per_second=0,5min_rate_reqs_per_second=0,75th_pc_request_time=0,95th_pc_request_time=0,999th_pc_request_time=0,99th_pc_request_time=0,avg_requests_per_second=0,avg_time_per_request=0,errors=0i,handler_start=1474662050861i,median_request_time=0,requests=0i,timeouts=0i,total_time=0 1478214949000000000
|
||||
> solr_queryhandler,core=main,handler=/tvrh,host=testhost 15min_rate_reqs_per_second=0,5min_rate_reqs_per_second=0,75th_pc_request_time=0,95th_pc_request_time=0,999th_pc_request_time=0,99th_pc_request_time=0,avg_requests_per_second=0,avg_time_per_request=0,errors=0i,handler_start=0i,median_request_time=0,requests=0i,timeouts=0i,total_time=0 1478214949000000000
|
||||
[…]
|
||||
```
|
|
@ -0,0 +1,442 @@
|
|||
package solr
|
||||
|
||||
import (
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"net/http"
|
||||
"strconv"
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
"github.com/influxdata/telegraf"
|
||||
"github.com/influxdata/telegraf/internal"
|
||||
"github.com/influxdata/telegraf/plugins/inputs"
|
||||
)
|
||||
|
||||
const mbeansPath = "/admin/mbeans?stats=true&wt=json&cat=CORE&cat=QUERYHANDLER&cat=UPDATEHANDLER&cat=CACHE"
|
||||
const adminCoresPath = "/solr/admin/cores?action=STATUS&wt=json"
|
||||
|
||||
type node struct {
|
||||
Host string `json:"host"`
|
||||
}
|
||||
|
||||
const sampleConfig = `
|
||||
## specify a list of one or more Solr servers
|
||||
servers = ["http://localhost:8983"]
|
||||
|
||||
## specify a list of one or more Solr cores (default - all)
|
||||
# cores = ["main"]
|
||||
`
|
||||
|
||||
// Solr is a plugin to read stats from one or many Solr servers
|
||||
type Solr struct {
|
||||
Local bool
|
||||
Servers []string
|
||||
HTTPTimeout internal.Duration
|
||||
Cores []string
|
||||
client *http.Client
|
||||
}
|
||||
|
||||
// AdminCoresStatus is an exported type that
|
||||
// contains a response with information about Solr cores.
|
||||
type AdminCoresStatus struct {
|
||||
Status map[string]struct {
|
||||
Index struct {
|
||||
SizeInBytes int64 `json:"sizeInBytes"`
|
||||
NumDocs int64 `json:"numDocs"`
|
||||
MaxDoc int64 `json:"maxDoc"`
|
||||
DeletedDocs int64 `json:"deletedDocs"`
|
||||
} `json:"index"`
|
||||
} `json:"status"`
|
||||
}
|
||||
|
||||
// MBeansData is an exported type that
|
||||
// contains a response from Solr with metrics
|
||||
type MBeansData struct {
|
||||
Headers ResponseHeader `json:"responseHeader"`
|
||||
SolrMbeans []json.RawMessage `json:"solr-mbeans"`
|
||||
}
|
||||
|
||||
// ResponseHeader is an exported type that
|
||||
// contains a response metrics: QTime and Status
|
||||
type ResponseHeader struct {
|
||||
QTime int64 `json:"QTime"`
|
||||
Status int64 `json:"status"`
|
||||
}
|
||||
|
||||
// Core is an exported type that
|
||||
// contains Core metrics
|
||||
type Core struct {
|
||||
Stats struct {
|
||||
DeletedDocs int64 `json:"deletedDocs"`
|
||||
MaxDoc int64 `json:"maxDoc"`
|
||||
NumDocs int64 `json:"numDocs"`
|
||||
} `json:"stats"`
|
||||
}
|
||||
|
||||
// QueryHandler is an exported type that
|
||||
// contains query handler metrics
|
||||
type QueryHandler struct {
|
||||
Stats struct {
|
||||
One5minRateReqsPerSecond float64 `json:"15minRateReqsPerSecond"`
|
||||
FiveMinRateReqsPerSecond float64 `json:"5minRateReqsPerSecond"`
|
||||
Seven5thPcRequestTime float64 `json:"75thPcRequestTime"`
|
||||
Nine5thPcRequestTime float64 `json:"95thPcRequestTime"`
|
||||
Nine99thPcRequestTime float64 `json:"999thPcRequestTime"`
|
||||
Nine9thPcRequestTime float64 `json:"99thPcRequestTime"`
|
||||
AvgRequestsPerSecond float64 `json:"avgRequestsPerSecond"`
|
||||
AvgTimePerRequest float64 `json:"avgTimePerRequest"`
|
||||
Errors int64 `json:"errors"`
|
||||
HandlerStart int64 `json:"handlerStart"`
|
||||
MedianRequestTime float64 `json:"medianRequestTime"`
|
||||
Requests int64 `json:"requests"`
|
||||
Timeouts int64 `json:"timeouts"`
|
||||
TotalTime float64 `json:"totalTime"`
|
||||
} `json:"stats"`
|
||||
}
|
||||
|
||||
// UpdateHandler is an exported type that
|
||||
// contains update handler metrics
|
||||
type UpdateHandler struct {
|
||||
Stats struct {
|
||||
Adds int64 `json:"adds"`
|
||||
AutocommitMaxDocs int64 `json:"autocommit maxDocs"`
|
||||
AutocommitMaxTime string `json:"autocommit maxTime"`
|
||||
Autocommits int64 `json:"autocommits"`
|
||||
Commits int64 `json:"commits"`
|
||||
CumulativeAdds int64 `json:"cumulative_adds"`
|
||||
CumulativeDeletesByID int64 `json:"cumulative_deletesById"`
|
||||
CumulativeDeletesByQuery int64 `json:"cumulative_deletesByQuery"`
|
||||
CumulativeErrors int64 `json:"cumulative_errors"`
|
||||
DeletesByID int64 `json:"deletesById"`
|
||||
DeletesByQuery int64 `json:"deletesByQuery"`
|
||||
DocsPending int64 `json:"docsPending"`
|
||||
Errors int64 `json:"errors"`
|
||||
ExpungeDeletes int64 `json:"expungeDeletes"`
|
||||
Optimizes int64 `json:"optimizes"`
|
||||
Rollbacks int64 `json:"rollbacks"`
|
||||
SoftAutocommits int64 `json:"soft autocommits"`
|
||||
} `json:"stats"`
|
||||
}
|
||||
|
||||
// Hitratio is an helper interface
|
||||
// so we can later on convert it to float64
|
||||
type Hitratio interface{}
|
||||
|
||||
// Cache is an exported type that
|
||||
// contains cache metrics
|
||||
type Cache struct {
|
||||
Stats struct {
|
||||
CumulativeEvictions int64 `json:"cumulative_evictions"`
|
||||
CumulativeHitratio Hitratio `json:"cumulative_hitratio"`
|
||||
CumulativeHits int64 `json:"cumulative_hits"`
|
||||
CumulativeInserts int64 `json:"cumulative_inserts"`
|
||||
CumulativeLookups int64 `json:"cumulative_lookups"`
|
||||
Evictions int64 `json:"evictions"`
|
||||
Hitratio Hitratio `json:"hitratio"`
|
||||
Hits int64 `json:"hits"`
|
||||
Inserts int64 `json:"inserts"`
|
||||
Lookups int64 `json:"lookups"`
|
||||
Size int64 `json:"size"`
|
||||
WarmupTime int64 `json:"warmupTime"`
|
||||
} `json:"stats"`
|
||||
}
|
||||
|
||||
// NewSolr return a new instance of Solr
|
||||
func NewSolr() *Solr {
|
||||
return &Solr{
|
||||
HTTPTimeout: internal.Duration{Duration: time.Second * 5},
|
||||
}
|
||||
}
|
||||
|
||||
// SampleConfig returns sample configuration for this plugin.
|
||||
func (s *Solr) SampleConfig() string {
|
||||
return sampleConfig
|
||||
}
|
||||
|
||||
// Description returns the plugin description.
|
||||
func (s *Solr) Description() string {
|
||||
return "Read stats from one or more Solr servers or cores"
|
||||
}
|
||||
|
||||
// Gather reads the stats from Solr and writes it to the
|
||||
// Accumulator.
|
||||
func (s *Solr) Gather(acc telegraf.Accumulator) error {
|
||||
if s.client == nil {
|
||||
client := s.createHTTPClient()
|
||||
s.client = client
|
||||
}
|
||||
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(len(s.Servers))
|
||||
|
||||
for _, serv := range s.Servers {
|
||||
go func(serv string, acc telegraf.Accumulator) {
|
||||
defer wg.Done()
|
||||
acc.AddError(s.gatherServerMetrics(serv, acc))
|
||||
}(serv, acc)
|
||||
}
|
||||
wg.Wait()
|
||||
return nil
|
||||
}
|
||||
|
||||
// Gather all metrics from server
|
||||
func (s *Solr) gatherServerMetrics(server string, acc telegraf.Accumulator) error {
|
||||
measurementTime := time.Now()
|
||||
adminCoresStatus := &AdminCoresStatus{}
|
||||
if err := s.gatherData(s.adminURL(server), adminCoresStatus); err != nil {
|
||||
return err
|
||||
}
|
||||
addAdminCoresStatusToAcc(acc, adminCoresStatus, measurementTime)
|
||||
cores := s.filterCores(getCoresFromStatus(adminCoresStatus))
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(len(cores))
|
||||
for _, core := range cores {
|
||||
go func(server string, core string, acc telegraf.Accumulator) {
|
||||
defer wg.Done()
|
||||
mBeansData := &MBeansData{}
|
||||
acc.AddError(s.gatherData(s.mbeansURL(server, core), mBeansData))
|
||||
acc.AddError(addCoreMetricsToAcc(acc, core, mBeansData, measurementTime))
|
||||
acc.AddError(addQueryHandlerMetricsToAcc(acc, core, mBeansData, measurementTime))
|
||||
acc.AddError(addUpdateHandlerMetricsToAcc(acc, core, mBeansData, measurementTime))
|
||||
acc.AddError(addCacheMetricsToAcc(acc, core, mBeansData, measurementTime))
|
||||
}(server, core, acc)
|
||||
}
|
||||
wg.Wait()
|
||||
return nil
|
||||
}
|
||||
|
||||
// Use cores from configuration if exists, else use cores from server
|
||||
func (s *Solr) filterCores(serverCores []string) []string {
|
||||
if len(s.Cores) == 0 {
|
||||
return serverCores
|
||||
}
|
||||
return s.Cores
|
||||
}
|
||||
|
||||
// Return list of cores from solr server
|
||||
func getCoresFromStatus(adminCoresStatus *AdminCoresStatus) []string {
|
||||
serverCores := []string{}
|
||||
for coreName := range adminCoresStatus.Status {
|
||||
serverCores = append(serverCores, coreName)
|
||||
}
|
||||
return serverCores
|
||||
}
|
||||
|
||||
// Add core metrics from admin to accumulator
|
||||
// This is the only point where size_in_bytes is available (as far as I checked)
|
||||
func addAdminCoresStatusToAcc(acc telegraf.Accumulator, adminCoreStatus *AdminCoresStatus, time time.Time) {
|
||||
for core, metrics := range adminCoreStatus.Status {
|
||||
coreFields := map[string]interface{}{
|
||||
"deleted_docs": metrics.Index.DeletedDocs,
|
||||
"max_docs": metrics.Index.MaxDoc,
|
||||
"num_docs": metrics.Index.NumDocs,
|
||||
"size_in_bytes": metrics.Index.SizeInBytes,
|
||||
}
|
||||
acc.AddFields(
|
||||
"solr_admin",
|
||||
coreFields,
|
||||
map[string]string{"core": core},
|
||||
time,
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
// Add core metrics section to accumulator
|
||||
func addCoreMetricsToAcc(acc telegraf.Accumulator, core string, mBeansData *MBeansData, time time.Time) error {
|
||||
var coreMetrics map[string]Core
|
||||
if err := json.Unmarshal(mBeansData.SolrMbeans[1], &coreMetrics); err != nil {
|
||||
return err
|
||||
}
|
||||
for name, metrics := range coreMetrics {
|
||||
if strings.Contains(name, "@") {
|
||||
continue
|
||||
}
|
||||
coreFields := map[string]interface{}{
|
||||
"deleted_docs": metrics.Stats.DeletedDocs,
|
||||
"max_docs": metrics.Stats.MaxDoc,
|
||||
"num_docs": metrics.Stats.NumDocs,
|
||||
}
|
||||
acc.AddFields(
|
||||
"solr_core",
|
||||
coreFields,
|
||||
map[string]string{
|
||||
"core": core,
|
||||
"handler": name},
|
||||
time,
|
||||
)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Add query metrics section to accumulator
|
||||
func addQueryHandlerMetricsToAcc(acc telegraf.Accumulator, core string, mBeansData *MBeansData, time time.Time) error {
|
||||
var queryMetrics map[string]QueryHandler
|
||||
|
||||
if err := json.Unmarshal(mBeansData.SolrMbeans[3], &queryMetrics); err != nil {
|
||||
return err
|
||||
}
|
||||
for name, metrics := range queryMetrics {
|
||||
coreFields := map[string]interface{}{
|
||||
"15min_rate_reqs_per_second": metrics.Stats.One5minRateReqsPerSecond,
|
||||
"5min_rate_reqs_per_second": metrics.Stats.FiveMinRateReqsPerSecond,
|
||||
"75th_pc_request_time": metrics.Stats.Seven5thPcRequestTime,
|
||||
"95th_pc_request_time": metrics.Stats.Nine5thPcRequestTime,
|
||||
"999th_pc_request_time": metrics.Stats.Nine99thPcRequestTime,
|
||||
"99th_pc_request_time": metrics.Stats.Nine9thPcRequestTime,
|
||||
"avg_requests_per_second": metrics.Stats.AvgRequestsPerSecond,
|
||||
"avg_time_per_request": metrics.Stats.AvgTimePerRequest,
|
||||
"errors": metrics.Stats.Errors,
|
||||
"handler_start": metrics.Stats.HandlerStart,
|
||||
"median_request_time": metrics.Stats.MedianRequestTime,
|
||||
"requests": metrics.Stats.Requests,
|
||||
"timeouts": metrics.Stats.Timeouts,
|
||||
"total_time": metrics.Stats.TotalTime,
|
||||
}
|
||||
acc.AddFields(
|
||||
"solr_queryhandler",
|
||||
coreFields,
|
||||
map[string]string{
|
||||
"core": core,
|
||||
"handler": name},
|
||||
time,
|
||||
)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Add update metrics section to accumulator
|
||||
func addUpdateHandlerMetricsToAcc(acc telegraf.Accumulator, core string, mBeansData *MBeansData, time time.Time) error {
|
||||
var updateMetrics map[string]UpdateHandler
|
||||
|
||||
if err := json.Unmarshal(mBeansData.SolrMbeans[5], &updateMetrics); err != nil {
|
||||
return err
|
||||
}
|
||||
for name, metrics := range updateMetrics {
|
||||
var autoCommitMaxTime int64
|
||||
if len(metrics.Stats.AutocommitMaxTime) > 2 {
|
||||
autoCommitMaxTime, _ = strconv.ParseInt(metrics.Stats.AutocommitMaxTime[:len(metrics.Stats.AutocommitMaxTime)-2], 0, 64)
|
||||
}
|
||||
coreFields := map[string]interface{}{
|
||||
"adds": metrics.Stats.Adds,
|
||||
"autocommit_max_docs": metrics.Stats.AutocommitMaxDocs,
|
||||
"autocommit_max_time": autoCommitMaxTime,
|
||||
"autocommits": metrics.Stats.Autocommits,
|
||||
"commits": metrics.Stats.Commits,
|
||||
"cumulative_adds": metrics.Stats.CumulativeAdds,
|
||||
"cumulative_deletes_by_id": metrics.Stats.CumulativeDeletesByID,
|
||||
"cumulative_deletes_by_query": metrics.Stats.CumulativeDeletesByQuery,
|
||||
"cumulative_errors": metrics.Stats.CumulativeErrors,
|
||||
"deletes_by_id": metrics.Stats.DeletesByID,
|
||||
"deletes_by_query": metrics.Stats.DeletesByQuery,
|
||||
"docs_pending": metrics.Stats.DocsPending,
|
||||
"errors": metrics.Stats.Errors,
|
||||
"expunge_deletes": metrics.Stats.ExpungeDeletes,
|
||||
"optimizes": metrics.Stats.Optimizes,
|
||||
"rollbacks": metrics.Stats.Rollbacks,
|
||||
"soft_autocommits": metrics.Stats.SoftAutocommits,
|
||||
}
|
||||
acc.AddFields(
|
||||
"solr_updatehandler",
|
||||
coreFields,
|
||||
map[string]string{
|
||||
"core": core,
|
||||
"handler": name},
|
||||
time,
|
||||
)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Get float64 from interface
|
||||
func getFloat(unk interface{}) float64 {
|
||||
switch i := unk.(type) {
|
||||
case float64:
|
||||
return i
|
||||
case string:
|
||||
f, _ := strconv.ParseFloat(i, 64)
|
||||
return f
|
||||
default:
|
||||
return float64(0)
|
||||
}
|
||||
}
|
||||
|
||||
// Add cache metrics section to accumulator
|
||||
func addCacheMetricsToAcc(acc telegraf.Accumulator, core string, mBeansData *MBeansData, time time.Time) error {
|
||||
var cacheMetrics map[string]Cache
|
||||
if err := json.Unmarshal(mBeansData.SolrMbeans[7], &cacheMetrics); err != nil {
|
||||
return err
|
||||
}
|
||||
for name, metrics := range cacheMetrics {
|
||||
cumulativeHits := getFloat(metrics.Stats.CumulativeHitratio)
|
||||
hitratio := getFloat(metrics.Stats.Hitratio)
|
||||
coreFields := map[string]interface{}{
|
||||
"cumulative_evictions": metrics.Stats.CumulativeEvictions,
|
||||
"cumulative_hitratio": cumulativeHits,
|
||||
"cumulative_hits": metrics.Stats.CumulativeHits,
|
||||
"cumulative_inserts": metrics.Stats.CumulativeInserts,
|
||||
"cumulative_lookups": metrics.Stats.CumulativeLookups,
|
||||
"evictions": metrics.Stats.Evictions,
|
||||
"hitratio": hitratio,
|
||||
"hits": metrics.Stats.Hits,
|
||||
"inserts": metrics.Stats.Inserts,
|
||||
"lookups": metrics.Stats.Lookups,
|
||||
"size": metrics.Stats.Size,
|
||||
"warmup_time": metrics.Stats.WarmupTime,
|
||||
}
|
||||
acc.AddFields(
|
||||
"solr_cache",
|
||||
coreFields,
|
||||
map[string]string{
|
||||
"core": core,
|
||||
"handler": name},
|
||||
time,
|
||||
)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Provide admin url
|
||||
func (s *Solr) adminURL(server string) string {
|
||||
return fmt.Sprintf("%s%s", server, adminCoresPath)
|
||||
}
|
||||
|
||||
// Provide mbeans url
|
||||
func (s *Solr) mbeansURL(server string, core string) string {
|
||||
return fmt.Sprintf("%s/solr/%s%s", server, core, mbeansPath)
|
||||
}
|
||||
|
||||
func (s *Solr) createHTTPClient() *http.Client {
|
||||
tr := &http.Transport{
|
||||
ResponseHeaderTimeout: s.HTTPTimeout.Duration,
|
||||
}
|
||||
client := &http.Client{
|
||||
Transport: tr,
|
||||
Timeout: s.HTTPTimeout.Duration,
|
||||
}
|
||||
|
||||
return client
|
||||
}
|
||||
|
||||
func (s *Solr) gatherData(url string, v interface{}) error {
|
||||
r, err := s.client.Get(url)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
defer r.Body.Close()
|
||||
if r.StatusCode != http.StatusOK {
|
||||
return fmt.Errorf("solr: API responded with status-code %d, expected %d, url %s",
|
||||
r.StatusCode, http.StatusOK, url)
|
||||
}
|
||||
if err = json.NewDecoder(r.Body).Decode(v); err != nil {
|
||||
return err
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func init() {
|
||||
inputs.Add("solr", func() telegraf.Input {
|
||||
return NewSolr()
|
||||
})
|
||||
}
|
|
@ -0,0 +1,62 @@
|
|||
package solr
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"net/http"
|
||||
"net/http/httptest"
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
"github.com/influxdata/telegraf/testutil"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func TestGatherStats(t *testing.T) {
|
||||
ts := createMockServer()
|
||||
solr := NewSolr()
|
||||
solr.Servers = []string{ts.URL}
|
||||
var acc testutil.Accumulator
|
||||
require.NoError(t, solr.Gather(&acc))
|
||||
|
||||
acc.AssertContainsTaggedFields(t, "solr_admin",
|
||||
solrAdminMainCoreStatusExpected,
|
||||
map[string]string{"core": "main"})
|
||||
|
||||
acc.AssertContainsTaggedFields(t, "solr_admin",
|
||||
solrAdminCore1StatusExpected,
|
||||
map[string]string{"core": "core1"})
|
||||
|
||||
acc.AssertContainsTaggedFields(t, "solr_core",
|
||||
solrCoreExpected,
|
||||
map[string]string{"core": "main", "handler": "searcher"})
|
||||
|
||||
acc.AssertContainsTaggedFields(t, "solr_queryhandler",
|
||||
solrQueryHandlerExpected,
|
||||
map[string]string{"core": "main", "handler": "org.apache.solr.handler.component.SearchHandler"})
|
||||
|
||||
acc.AssertContainsTaggedFields(t, "solr_updatehandler",
|
||||
solrUpdateHandlerExpected,
|
||||
map[string]string{"core": "main", "handler": "updateHandler"})
|
||||
|
||||
acc.AssertContainsTaggedFields(t, "solr_cache",
|
||||
solrCacheExpected,
|
||||
map[string]string{"core": "main", "handler": "filterCache"})
|
||||
}
|
||||
|
||||
func createMockServer() *httptest.Server {
|
||||
return httptest.NewServer(http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {
|
||||
if strings.Contains(r.URL.Path, "/solr/admin/cores") {
|
||||
w.WriteHeader(http.StatusOK)
|
||||
fmt.Fprintln(w, statusResponse)
|
||||
} else if strings.Contains(r.URL.Path, "solr/main/admin") {
|
||||
w.WriteHeader(http.StatusOK)
|
||||
fmt.Fprintln(w, mBeansMainResponse)
|
||||
} else if strings.Contains(r.URL.Path, "solr/core1/admin") {
|
||||
w.WriteHeader(http.StatusOK)
|
||||
fmt.Fprintln(w, mBeansCore1Response)
|
||||
} else {
|
||||
w.WriteHeader(http.StatusNotFound)
|
||||
fmt.Fprintln(w, "nope")
|
||||
}
|
||||
}))
|
||||
}
|
File diff suppressed because it is too large
Load Diff
Loading…
Reference in New Issue