mirror of
https://github.com/grafana/grafana.git
synced 2025-02-14 17:43:35 -06:00
* Convert Cloud Monitoring (Stackdriver) Datasource to use Dataframes #29830 * add deeplink into config * omggggggggggggggg this deeplink works! * move unit to the backend part * remove unit from frontend * only set the config fields[1] for deeplink and unit * refactory + fix some test * remove frontend test for unit * adding backend test for unit mapping * resolve review * rewrtie unit logic to do exactly the same as frontend filter * refactory
900 lines
27 KiB
Go
900 lines
27 KiB
Go
package cloudmonitoring
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"math"
|
|
"net/http"
|
|
"net/url"
|
|
"path"
|
|
"regexp"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/grafana/grafana-plugin-sdk-go/data"
|
|
"github.com/grafana/grafana/pkg/api/pluginproxy"
|
|
"github.com/grafana/grafana/pkg/components/simplejson"
|
|
"github.com/grafana/grafana/pkg/infra/log"
|
|
"github.com/grafana/grafana/pkg/models"
|
|
"github.com/grafana/grafana/pkg/plugins"
|
|
"github.com/grafana/grafana/pkg/setting"
|
|
"github.com/grafana/grafana/pkg/tsdb"
|
|
"github.com/grafana/grafana/pkg/tsdb/sqleng"
|
|
"github.com/opentracing/opentracing-go"
|
|
"golang.org/x/net/context/ctxhttp"
|
|
"golang.org/x/oauth2/google"
|
|
)
|
|
|
|
var (
|
|
slog log.Logger
|
|
)
|
|
|
|
var (
|
|
matchAllCap = regexp.MustCompile("(.)([A-Z][a-z]*)")
|
|
legendKeyFormat = regexp.MustCompile(`\{\{\s*(.+?)\s*\}\}`)
|
|
metricNameFormat = regexp.MustCompile(`([\w\d_]+)\.(googleapis\.com|io)/(.+)`)
|
|
wildcardRegexRe = regexp.MustCompile(`[-\/^$+?.()|[\]{}]`)
|
|
alignmentPeriodRe = regexp.MustCompile("[0-9]+")
|
|
cloudMonitoringUnitMappings = map[string]string{
|
|
"bit": "bits",
|
|
"By": "bytes",
|
|
"s": "s",
|
|
"min": "m",
|
|
"h": "h",
|
|
"d": "d",
|
|
"us": "µs",
|
|
"ms": "ms",
|
|
"ns": "ns",
|
|
"percent": "percent",
|
|
"MiBy": "mbytes",
|
|
"By/s": "Bps",
|
|
"GBy": "decgbytes",
|
|
}
|
|
)
|
|
|
|
const (
|
|
gceAuthentication string = "gce"
|
|
jwtAuthentication string = "jwt"
|
|
metricQueryType string = "metrics"
|
|
sloQueryType string = "slo"
|
|
)
|
|
|
|
// CloudMonitoringExecutor executes queries for the CloudMonitoring datasource
|
|
type CloudMonitoringExecutor struct {
|
|
httpClient *http.Client
|
|
dsInfo *models.DataSource
|
|
}
|
|
|
|
// NewCloudMonitoringExecutor initializes a http client
|
|
func NewCloudMonitoringExecutor(dsInfo *models.DataSource) (tsdb.TsdbQueryEndpoint, error) {
|
|
httpClient, err := dsInfo.GetHttpClient()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &CloudMonitoringExecutor{
|
|
httpClient: httpClient,
|
|
dsInfo: dsInfo,
|
|
}, nil
|
|
}
|
|
|
|
func init() {
|
|
slog = log.New("tsdb.cloudMonitoring")
|
|
tsdb.RegisterTsdbQueryEndpoint("stackdriver", NewCloudMonitoringExecutor)
|
|
}
|
|
|
|
// Query takes in the frontend queries, parses them into the CloudMonitoring query format
|
|
// executes the queries against the CloudMonitoring API and parses the response into
|
|
// the time series or table format
|
|
func (e *CloudMonitoringExecutor) Query(ctx context.Context, dsInfo *models.DataSource, tsdbQuery *tsdb.TsdbQuery) (*tsdb.Response, error) {
|
|
var result *tsdb.Response
|
|
var err error
|
|
queryType := tsdbQuery.Queries[0].Model.Get("type").MustString("")
|
|
|
|
switch queryType {
|
|
case "annotationQuery":
|
|
result, err = e.executeAnnotationQuery(ctx, tsdbQuery)
|
|
case "getGCEDefaultProject":
|
|
result, err = e.getGCEDefaultProject(ctx, tsdbQuery)
|
|
case "timeSeriesQuery":
|
|
fallthrough
|
|
default:
|
|
result, err = e.executeTimeSeriesQuery(ctx, tsdbQuery)
|
|
}
|
|
|
|
return result, err
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) getGCEDefaultProject(ctx context.Context, tsdbQuery *tsdb.TsdbQuery) (*tsdb.Response, error) {
|
|
result := &tsdb.Response{
|
|
Results: make(map[string]*tsdb.QueryResult),
|
|
}
|
|
refId := tsdbQuery.Queries[0].RefId
|
|
queryResult := &tsdb.QueryResult{Meta: simplejson.New(), RefId: refId}
|
|
|
|
gceDefaultProject, err := e.getDefaultProject(ctx)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to retrieve default project from GCE metadata server, error: %w", err)
|
|
}
|
|
|
|
queryResult.Meta.Set("defaultProject", gceDefaultProject)
|
|
result.Results[refId] = queryResult
|
|
|
|
return result, nil
|
|
}
|
|
|
|
func (query *cloudMonitoringQuery) isSLO() bool {
|
|
return query.Slo != ""
|
|
}
|
|
|
|
func (query *cloudMonitoringQuery) buildDeepLink() string {
|
|
if query.isSLO() {
|
|
return ""
|
|
}
|
|
|
|
filter := query.Params.Get("filter")
|
|
if !strings.Contains(filter, "resource.type=") {
|
|
resourceType := query.Params.Get("resourceType")
|
|
if resourceType == "" {
|
|
slog.Error("Failed to generate deep link: no resource type found", "ProjectName", query.ProjectName, "query", query.RefID)
|
|
return ""
|
|
}
|
|
filter = fmt.Sprintf(`resource.type="%s" %s`, resourceType, filter)
|
|
}
|
|
|
|
u, err := url.Parse("https://console.cloud.google.com/monitoring/metrics-explorer")
|
|
if err != nil {
|
|
slog.Error("Failed to generate deep link: unable to parse metrics explorer URL", "ProjectName", query.ProjectName, "query", query.RefID)
|
|
return ""
|
|
}
|
|
|
|
q := u.Query()
|
|
q.Set("project", query.ProjectName)
|
|
q.Set("Grafana_deeplink", "true")
|
|
|
|
pageState := map[string]interface{}{
|
|
"xyChart": map[string]interface{}{
|
|
"constantLines": []string{},
|
|
"dataSets": []map[string]interface{}{
|
|
{
|
|
"timeSeriesFilter": map[string]interface{}{
|
|
"aggregations": []string{},
|
|
"crossSeriesReducer": query.Params.Get("aggregation.crossSeriesReducer"),
|
|
"filter": filter,
|
|
"groupByFields": query.Params["aggregation.groupByFields"],
|
|
"minAlignmentPeriod": strings.TrimPrefix(query.Params.Get("aggregation.alignmentPeriod"), "+"), // get rid of leading +
|
|
"perSeriesAligner": query.Params.Get("aggregation.perSeriesAligner"),
|
|
"secondaryGroupByFields": []string{},
|
|
"unitOverride": "1",
|
|
},
|
|
},
|
|
},
|
|
"timeshiftDuration": "0s",
|
|
"y1Axis": map[string]string{
|
|
"label": "y1Axis",
|
|
"scale": "LINEAR",
|
|
},
|
|
},
|
|
"timeSelection": map[string]string{
|
|
"timeRange": "custom",
|
|
"start": query.Params.Get("interval.startTime"),
|
|
"end": query.Params.Get("interval.endTime"),
|
|
},
|
|
}
|
|
|
|
blob, err := json.Marshal(pageState)
|
|
if err != nil {
|
|
slog.Error("Failed to generate deep link", "pageState", pageState, "ProjectName", query.ProjectName, "query", query.RefID)
|
|
return ""
|
|
}
|
|
|
|
q.Set("pageState", string(blob))
|
|
u.RawQuery = q.Encode()
|
|
|
|
accountChooserURL, err := url.Parse("https://accounts.google.com/AccountChooser")
|
|
if err != nil {
|
|
slog.Error("Failed to generate deep link: unable to parse account chooser URL", "ProjectName", query.ProjectName, "query", query.RefID)
|
|
return ""
|
|
}
|
|
accountChooserQuery := accountChooserURL.Query()
|
|
accountChooserQuery.Set("continue", u.String())
|
|
accountChooserURL.RawQuery = accountChooserQuery.Encode()
|
|
|
|
return accountChooserURL.String()
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) executeTimeSeriesQuery(ctx context.Context, tsdbQuery *tsdb.TsdbQuery) (*tsdb.Response, error) {
|
|
result := &tsdb.Response{
|
|
Results: make(map[string]*tsdb.QueryResult),
|
|
}
|
|
|
|
queries, err := e.buildQueries(tsdbQuery)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
unit := e.resolvePanelUnitFromQueries(queries)
|
|
|
|
for _, query := range queries {
|
|
queryRes, resp, err := e.executeQuery(ctx, query, tsdbQuery)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
resourceType := ""
|
|
for _, s := range resp.TimeSeries {
|
|
resourceType = s.Resource.Type
|
|
// set the first resource type found
|
|
break
|
|
}
|
|
query.Params.Set("resourceType", resourceType)
|
|
|
|
err = e.parseResponse(queryRes, resp, query)
|
|
if err != nil {
|
|
queryRes.Error = err
|
|
}
|
|
|
|
if len(unit) > 0 {
|
|
frames, _ := queryRes.Dataframes.Decoded()
|
|
for i := range frames {
|
|
if frames[i].Fields[1].Config == nil {
|
|
frames[i].Fields[1].Config = &data.FieldConfig{}
|
|
}
|
|
frames[i].Fields[1].Config.Unit = unit
|
|
}
|
|
queryRes.Dataframes = tsdb.NewDecodedDataFrames(frames)
|
|
}
|
|
result.Results[query.RefID] = queryRes
|
|
}
|
|
|
|
return result, nil
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) resolvePanelUnitFromQueries(queries []*cloudMonitoringQuery) string {
|
|
if len(queries) == 0 {
|
|
return ""
|
|
}
|
|
unit := queries[0].Unit
|
|
if len(queries) > 1 {
|
|
for _, query := range queries[1:] {
|
|
if query.Unit != unit {
|
|
return ""
|
|
}
|
|
}
|
|
}
|
|
if len(unit) > 0 {
|
|
if val, ok := cloudMonitoringUnitMappings[unit]; ok {
|
|
return val
|
|
}
|
|
}
|
|
return ""
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) buildQueries(tsdbQuery *tsdb.TsdbQuery) ([]*cloudMonitoringQuery, error) {
|
|
cloudMonitoringQueries := []*cloudMonitoringQuery{}
|
|
|
|
startTime, err := tsdbQuery.TimeRange.ParseFrom()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
endTime, err := tsdbQuery.TimeRange.ParseTo()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
durationSeconds := int(endTime.Sub(startTime).Seconds())
|
|
|
|
for _, query := range tsdbQuery.Queries {
|
|
migrateLegacyQueryModel(query)
|
|
q := grafanaQuery{}
|
|
model, _ := query.Model.MarshalJSON()
|
|
if err := json.Unmarshal(model, &q); err != nil {
|
|
return nil, fmt.Errorf("could not unmarshal CloudMonitoringQuery json: %w", err)
|
|
}
|
|
var target string
|
|
params := url.Values{}
|
|
params.Add("interval.startTime", startTime.UTC().Format(time.RFC3339))
|
|
params.Add("interval.endTime", endTime.UTC().Format(time.RFC3339))
|
|
|
|
sq := &cloudMonitoringQuery{
|
|
RefID: query.RefId,
|
|
GroupBys: []string{},
|
|
}
|
|
|
|
if q.QueryType == metricQueryType {
|
|
sq.AliasBy = q.MetricQuery.AliasBy
|
|
sq.GroupBys = append(sq.GroupBys, q.MetricQuery.GroupBys...)
|
|
sq.ProjectName = q.MetricQuery.ProjectName
|
|
if q.MetricQuery.View == "" {
|
|
q.MetricQuery.View = "FULL"
|
|
}
|
|
params.Add("filter", buildFilterString(q.MetricQuery.MetricType, q.MetricQuery.Filters))
|
|
params.Add("view", q.MetricQuery.View)
|
|
setMetricAggParams(¶ms, &q.MetricQuery, durationSeconds, query.IntervalMs)
|
|
} else if q.QueryType == sloQueryType {
|
|
sq.AliasBy = q.SloQuery.AliasBy
|
|
sq.ProjectName = q.SloQuery.ProjectName
|
|
sq.Selector = q.SloQuery.SelectorName
|
|
sq.Service = q.SloQuery.ServiceId
|
|
sq.Slo = q.SloQuery.SloId
|
|
params.Add("filter", buildSLOFilterExpression(q.SloQuery))
|
|
setSloAggParams(¶ms, &q.SloQuery, durationSeconds, query.IntervalMs)
|
|
}
|
|
|
|
target = params.Encode()
|
|
sq.Target = target
|
|
sq.Params = params
|
|
sq.Unit = q.MetricQuery.Unit
|
|
if setting.Env == setting.Dev {
|
|
slog.Debug("CloudMonitoring request", "params", params)
|
|
}
|
|
|
|
cloudMonitoringQueries = append(cloudMonitoringQueries, sq)
|
|
}
|
|
|
|
return cloudMonitoringQueries, nil
|
|
}
|
|
|
|
func migrateLegacyQueryModel(query *tsdb.Query) {
|
|
mq := query.Model.Get("metricQuery").MustMap()
|
|
if mq == nil {
|
|
migratedModel := simplejson.NewFromAny(map[string]interface{}{
|
|
"queryType": metricQueryType,
|
|
"metricQuery": query.Model.MustMap(),
|
|
})
|
|
query.Model = migratedModel
|
|
}
|
|
}
|
|
|
|
func reverse(s string) string {
|
|
chars := []rune(s)
|
|
for i, j := 0, len(chars)-1; i < j; i, j = i+1, j-1 {
|
|
chars[i], chars[j] = chars[j], chars[i]
|
|
}
|
|
return string(chars)
|
|
}
|
|
|
|
func interpolateFilterWildcards(value string) string {
|
|
matches := strings.Count(value, "*")
|
|
switch {
|
|
case matches == 2 && strings.HasSuffix(value, "*") && strings.HasPrefix(value, "*"):
|
|
value = strings.ReplaceAll(value, "*", "")
|
|
value = fmt.Sprintf(`has_substring("%s")`, value)
|
|
case matches == 1 && strings.HasPrefix(value, "*"):
|
|
value = strings.Replace(value, "*", "", 1)
|
|
value = fmt.Sprintf(`ends_with("%s")`, value)
|
|
case matches == 1 && strings.HasSuffix(value, "*"):
|
|
value = reverse(strings.Replace(reverse(value), "*", "", 1))
|
|
value = fmt.Sprintf(`starts_with("%s")`, value)
|
|
case matches != 0:
|
|
value = string(wildcardRegexRe.ReplaceAllFunc([]byte(value), func(in []byte) []byte {
|
|
return []byte(strings.Replace(string(in), string(in), `\\`+string(in), 1))
|
|
}))
|
|
value = strings.ReplaceAll(value, "*", ".*")
|
|
value = strings.ReplaceAll(value, `"`, `\\"`)
|
|
value = fmt.Sprintf(`monitoring.regex.full_match("^%s$")`, value)
|
|
}
|
|
|
|
return value
|
|
}
|
|
|
|
func buildFilterString(metricType string, filterParts []string) string {
|
|
filterString := ""
|
|
for i, part := range filterParts {
|
|
mod := i % 4
|
|
switch {
|
|
case part == "AND":
|
|
filterString += " "
|
|
case mod == 2:
|
|
operator := filterParts[i-1]
|
|
switch {
|
|
case operator == "=~" || operator == "!=~":
|
|
filterString = reverse(strings.Replace(reverse(filterString), "~", "", 1))
|
|
filterString += fmt.Sprintf(`monitoring.regex.full_match("%s")`, part)
|
|
case strings.Contains(part, "*"):
|
|
filterString += interpolateFilterWildcards(part)
|
|
default:
|
|
filterString += fmt.Sprintf(`"%s"`, part)
|
|
}
|
|
default:
|
|
filterString += part
|
|
}
|
|
}
|
|
|
|
return strings.Trim(fmt.Sprintf(`metric.type="%s" %s`, metricType, filterString), " ")
|
|
}
|
|
|
|
func buildSLOFilterExpression(q sloQuery) string {
|
|
return fmt.Sprintf(`%s("projects/%s/services/%s/serviceLevelObjectives/%s")`, q.SelectorName, q.ProjectName, q.ServiceId, q.SloId)
|
|
}
|
|
|
|
func setMetricAggParams(params *url.Values, query *metricQuery, durationSeconds int, intervalMs int64) {
|
|
if query.CrossSeriesReducer == "" {
|
|
query.CrossSeriesReducer = "REDUCE_NONE"
|
|
}
|
|
|
|
if query.PerSeriesAligner == "" {
|
|
query.PerSeriesAligner = "ALIGN_MEAN"
|
|
}
|
|
|
|
params.Add("aggregation.crossSeriesReducer", query.CrossSeriesReducer)
|
|
params.Add("aggregation.perSeriesAligner", query.PerSeriesAligner)
|
|
params.Add("aggregation.alignmentPeriod", calculateAlignmentPeriod(query.AlignmentPeriod, intervalMs, durationSeconds))
|
|
|
|
for _, groupBy := range query.GroupBys {
|
|
params.Add("aggregation.groupByFields", groupBy)
|
|
}
|
|
}
|
|
|
|
func setSloAggParams(params *url.Values, query *sloQuery, durationSeconds int, intervalMs int64) {
|
|
params.Add("aggregation.alignmentPeriod", calculateAlignmentPeriod(query.AlignmentPeriod, intervalMs, durationSeconds))
|
|
if query.SelectorName == "select_slo_health" {
|
|
params.Add("aggregation.perSeriesAligner", "ALIGN_MEAN")
|
|
} else {
|
|
params.Add("aggregation.perSeriesAligner", "ALIGN_NEXT_OLDER")
|
|
}
|
|
}
|
|
|
|
func calculateAlignmentPeriod(alignmentPeriod string, intervalMs int64, durationSeconds int) string {
|
|
if alignmentPeriod == "grafana-auto" || alignmentPeriod == "" {
|
|
alignmentPeriodValue := int(math.Max(float64(intervalMs)/1000, 60.0))
|
|
alignmentPeriod = "+" + strconv.Itoa(alignmentPeriodValue) + "s"
|
|
}
|
|
|
|
if alignmentPeriod == "cloud-monitoring-auto" || alignmentPeriod == "stackdriver-auto" { // legacy
|
|
alignmentPeriodValue := int(math.Max(float64(durationSeconds), 60.0))
|
|
switch {
|
|
case alignmentPeriodValue < 60*60*23:
|
|
alignmentPeriod = "+60s"
|
|
case alignmentPeriodValue < 60*60*24*6:
|
|
alignmentPeriod = "+300s"
|
|
default:
|
|
alignmentPeriod = "+3600s"
|
|
}
|
|
}
|
|
|
|
return alignmentPeriod
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) executeQuery(ctx context.Context, query *cloudMonitoringQuery, tsdbQuery *tsdb.TsdbQuery) (*tsdb.QueryResult, cloudMonitoringResponse, error) {
|
|
queryResult := &tsdb.QueryResult{Meta: simplejson.New(), RefId: query.RefID}
|
|
projectName := query.ProjectName
|
|
if projectName == "" {
|
|
defaultProject, err := e.getDefaultProject(ctx)
|
|
if err != nil {
|
|
queryResult.Error = err
|
|
return queryResult, cloudMonitoringResponse{}, nil
|
|
}
|
|
projectName = defaultProject
|
|
slog.Info("No project name set on query, using project name from datasource", "projectName", projectName)
|
|
}
|
|
|
|
req, err := e.createRequest(ctx, e.dsInfo, query, fmt.Sprintf("cloudmonitoring%s", "v3/projects/"+projectName+"/timeSeries"))
|
|
if err != nil {
|
|
queryResult.Error = err
|
|
return queryResult, cloudMonitoringResponse{}, nil
|
|
}
|
|
|
|
req.URL.RawQuery = query.Params.Encode()
|
|
queryResult.Meta.Set(sqleng.MetaKeyExecutedQueryString, req.URL.RawQuery)
|
|
alignmentPeriod, ok := req.URL.Query()["aggregation.alignmentPeriod"]
|
|
|
|
if ok {
|
|
seconds, err := strconv.ParseInt(alignmentPeriodRe.FindString(alignmentPeriod[0]), 10, 64)
|
|
if err == nil {
|
|
queryResult.Meta.Set("alignmentPeriod", seconds)
|
|
}
|
|
}
|
|
|
|
span, ctx := opentracing.StartSpanFromContext(ctx, "cloudMonitoring query")
|
|
span.SetTag("target", query.Target)
|
|
span.SetTag("from", tsdbQuery.TimeRange.From)
|
|
span.SetTag("until", tsdbQuery.TimeRange.To)
|
|
span.SetTag("datasource_id", e.dsInfo.Id)
|
|
span.SetTag("org_id", e.dsInfo.OrgId)
|
|
|
|
defer span.Finish()
|
|
|
|
if err := opentracing.GlobalTracer().Inject(
|
|
span.Context(),
|
|
opentracing.HTTPHeaders,
|
|
opentracing.HTTPHeadersCarrier(req.Header)); err != nil {
|
|
queryResult.Error = err
|
|
return queryResult, cloudMonitoringResponse{}, nil
|
|
}
|
|
|
|
res, err := ctxhttp.Do(ctx, e.httpClient, req)
|
|
if err != nil {
|
|
queryResult.Error = err
|
|
return queryResult, cloudMonitoringResponse{}, nil
|
|
}
|
|
|
|
data, err := e.unmarshalResponse(res)
|
|
if err != nil {
|
|
queryResult.Error = err
|
|
return queryResult, cloudMonitoringResponse{}, nil
|
|
}
|
|
|
|
return queryResult, data, nil
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) unmarshalResponse(res *http.Response) (cloudMonitoringResponse, error) {
|
|
body, err := ioutil.ReadAll(res.Body)
|
|
if err != nil {
|
|
return cloudMonitoringResponse{}, err
|
|
}
|
|
defer func() {
|
|
if err := res.Body.Close(); err != nil {
|
|
slog.Warn("Failed to close response body", "err", err)
|
|
}
|
|
}()
|
|
|
|
if res.StatusCode/100 != 2 {
|
|
slog.Error("Request failed", "status", res.Status, "body", string(body))
|
|
return cloudMonitoringResponse{}, fmt.Errorf(string(body))
|
|
}
|
|
|
|
var data cloudMonitoringResponse
|
|
err = json.Unmarshal(body, &data)
|
|
if err != nil {
|
|
slog.Error("Failed to unmarshal CloudMonitoring response", "error", err, "status", res.Status, "body", string(body))
|
|
return cloudMonitoringResponse{}, err
|
|
}
|
|
|
|
return data, nil
|
|
}
|
|
|
|
func handleDistributionSeries(series timeSeries, defaultMetricName string, seriesLabels map[string]string,
|
|
query *cloudMonitoringQuery, queryRes *tsdb.QueryResult, frame *data.Frame) {
|
|
for i := 0; i < len(series.Points); i++ {
|
|
point := series.Points[i]
|
|
value := point.Value.DoubleValue
|
|
|
|
if series.ValueType == "INT64" {
|
|
parsedValue, err := strconv.ParseFloat(point.Value.IntValue, 64)
|
|
if err == nil {
|
|
value = parsedValue
|
|
}
|
|
}
|
|
|
|
if series.ValueType == "BOOL" {
|
|
if point.Value.BoolValue {
|
|
value = 1
|
|
} else {
|
|
value = 0
|
|
}
|
|
}
|
|
frame.SetRow(len(series.Points)-1-i, point.Interval.EndTime, value)
|
|
}
|
|
|
|
metricName := formatLegendKeys(series.Metric.Type, defaultMetricName, seriesLabels, nil, query)
|
|
dataField := frame.Fields[1]
|
|
dataField.Name = metricName
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) parseResponse(queryRes *tsdb.QueryResult, cmr cloudMonitoringResponse, query *cloudMonitoringQuery) error {
|
|
labels := make(map[string]map[string]bool)
|
|
frames := data.Frames{}
|
|
for _, series := range cmr.TimeSeries {
|
|
seriesLabels := data.Labels{}
|
|
defaultMetricName := series.Metric.Type
|
|
|
|
labels["resource.type"] = map[string]bool{series.Resource.Type: true}
|
|
seriesLabels["resource.type"] = series.Resource.Type
|
|
|
|
frame := data.NewFrameOfFieldTypes("", len(series.Points), data.FieldTypeTime, data.FieldTypeFloat64)
|
|
frame.RefID = query.RefID
|
|
|
|
for key, value := range series.Metric.Labels {
|
|
if _, ok := labels["metric.label."+key]; !ok {
|
|
labels["metric.label."+key] = map[string]bool{}
|
|
}
|
|
labels["metric.label."+key][value] = true
|
|
seriesLabels["metric.label."+key] = value
|
|
|
|
if len(query.GroupBys) == 0 || containsLabel(query.GroupBys, "metric.label."+key) {
|
|
defaultMetricName += " " + value
|
|
}
|
|
}
|
|
|
|
for key, value := range series.Resource.Labels {
|
|
if _, ok := labels["resource.label."+key]; !ok {
|
|
labels["resource.label."+key] = map[string]bool{}
|
|
}
|
|
labels["resource.label."+key][value] = true
|
|
seriesLabels["resource.label."+key] = value
|
|
|
|
if containsLabel(query.GroupBys, "resource.label."+key) {
|
|
defaultMetricName += " " + value
|
|
}
|
|
}
|
|
|
|
for labelType, labelTypeValues := range series.MetaData {
|
|
for labelKey, labelValue := range labelTypeValues {
|
|
key := toSnakeCase(fmt.Sprintf("metadata.%s.%s", labelType, labelKey))
|
|
if _, ok := labels[key]; !ok {
|
|
labels[key] = map[string]bool{}
|
|
}
|
|
|
|
switch v := labelValue.(type) {
|
|
case string:
|
|
labels[key][v] = true
|
|
seriesLabels[key] = v
|
|
case bool:
|
|
strVal := strconv.FormatBool(v)
|
|
labels[key][strVal] = true
|
|
seriesLabels[key] = strVal
|
|
case []interface{}:
|
|
for _, v := range v {
|
|
strVal := v.(string)
|
|
labels[key][strVal] = true
|
|
if len(seriesLabels[key]) > 0 {
|
|
strVal = fmt.Sprintf("%s, %s", seriesLabels[key], strVal)
|
|
}
|
|
seriesLabels[key] = strVal
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// reverse the order to be ascending
|
|
if series.ValueType != "DISTRIBUTION" {
|
|
handleDistributionSeries(
|
|
series, defaultMetricName, seriesLabels, query, queryRes, frame)
|
|
frames = append(frames, frame)
|
|
} else {
|
|
buckets := make(map[int]*data.Frame)
|
|
for i := len(series.Points) - 1; i >= 0; i-- {
|
|
point := series.Points[i]
|
|
if len(point.Value.DistributionValue.BucketCounts) == 0 {
|
|
continue
|
|
}
|
|
maxKey := 0
|
|
for i := 0; i < len(point.Value.DistributionValue.BucketCounts); i++ {
|
|
value, err := strconv.ParseFloat(point.Value.DistributionValue.BucketCounts[i], 64)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
if _, ok := buckets[i]; !ok {
|
|
// set lower bounds
|
|
// https://cloud.google.com/monitoring/api/ref_v3/rest/v3/TimeSeries#Distribution
|
|
bucketBound := calcBucketBound(point.Value.DistributionValue.BucketOptions, i)
|
|
additionalLabels := map[string]string{"bucket": bucketBound}
|
|
|
|
timeField := data.NewField(data.TimeSeriesTimeFieldName, nil, []time.Time{})
|
|
valueField := data.NewField(data.TimeSeriesValueFieldName, nil, []float64{})
|
|
|
|
frameName := formatLegendKeys(series.Metric.Type, defaultMetricName, nil, additionalLabels, query)
|
|
valueField.Name = frameName
|
|
buckets[i] = &data.Frame{
|
|
Name: frameName,
|
|
Fields: []*data.Field{
|
|
timeField,
|
|
valueField,
|
|
},
|
|
RefID: query.RefID,
|
|
}
|
|
|
|
if maxKey < i {
|
|
maxKey = i
|
|
}
|
|
}
|
|
buckets[i].AppendRow(point.Interval.EndTime, value)
|
|
}
|
|
for i := 0; i < maxKey; i++ {
|
|
if _, ok := buckets[i]; !ok {
|
|
bucketBound := calcBucketBound(point.Value.DistributionValue.BucketOptions, i)
|
|
additionalLabels := data.Labels{"bucket": bucketBound}
|
|
timeField := data.NewField(data.TimeSeriesTimeFieldName, nil, []time.Time{})
|
|
valueField := data.NewField(data.TimeSeriesValueFieldName, nil, []float64{})
|
|
frameName := formatLegendKeys(series.Metric.Type, defaultMetricName, seriesLabels, additionalLabels, query)
|
|
valueField.Name = frameName
|
|
buckets[i] = &data.Frame{
|
|
Name: frameName,
|
|
Fields: []*data.Field{
|
|
timeField,
|
|
valueField,
|
|
},
|
|
RefID: query.RefID,
|
|
}
|
|
}
|
|
}
|
|
}
|
|
for i := 0; i < len(buckets); i++ {
|
|
frames = append(frames, buckets[i])
|
|
}
|
|
}
|
|
}
|
|
if len(cmr.TimeSeries) > 0 {
|
|
frames = addConfigData(frames, query)
|
|
}
|
|
|
|
queryRes.Dataframes = tsdb.NewDecodedDataFrames(frames)
|
|
|
|
labelsByKey := make(map[string][]string)
|
|
for key, values := range labels {
|
|
for value := range values {
|
|
labelsByKey[key] = append(labelsByKey[key], value)
|
|
}
|
|
}
|
|
|
|
queryRes.Meta.Set("labels", labelsByKey)
|
|
queryRes.Meta.Set("groupBys", query.GroupBys)
|
|
return nil
|
|
}
|
|
|
|
func addConfigData(frames data.Frames, query *cloudMonitoringQuery) data.Frames {
|
|
dl := query.buildDeepLink()
|
|
for i := range frames {
|
|
if frames[i].Fields[1].Config == nil {
|
|
frames[i].Fields[1].Config = &data.FieldConfig{}
|
|
}
|
|
deepLink := data.DataLink{
|
|
Title: "View in Metrics Explorer",
|
|
TargetBlank: true,
|
|
URL: dl,
|
|
}
|
|
frames[i].Fields[1].Config.Links = append(frames[i].Fields[1].Config.Links, deepLink)
|
|
}
|
|
return frames
|
|
}
|
|
|
|
func toSnakeCase(str string) string {
|
|
return strings.ToLower(matchAllCap.ReplaceAllString(str, "${1}_${2}"))
|
|
}
|
|
|
|
func containsLabel(labels []string, newLabel string) bool {
|
|
for _, val := range labels {
|
|
if val == newLabel {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func formatLegendKeys(metricType string, defaultMetricName string, labels map[string]string, additionalLabels map[string]string, query *cloudMonitoringQuery) string {
|
|
if query.AliasBy == "" {
|
|
return defaultMetricName
|
|
}
|
|
|
|
result := legendKeyFormat.ReplaceAllFunc([]byte(query.AliasBy), func(in []byte) []byte {
|
|
metaPartName := strings.Replace(string(in), "{{", "", 1)
|
|
metaPartName = strings.Replace(metaPartName, "}}", "", 1)
|
|
metaPartName = strings.TrimSpace(metaPartName)
|
|
|
|
if metaPartName == "metric.type" {
|
|
return []byte(metricType)
|
|
}
|
|
|
|
metricPart := replaceWithMetricPart(metaPartName, metricType)
|
|
|
|
if metricPart != nil {
|
|
return metricPart
|
|
}
|
|
|
|
if val, exists := labels[metaPartName]; exists {
|
|
return []byte(val)
|
|
}
|
|
|
|
if val, exists := additionalLabels[metaPartName]; exists {
|
|
return []byte(val)
|
|
}
|
|
|
|
if metaPartName == "project" && query.ProjectName != "" {
|
|
return []byte(query.ProjectName)
|
|
}
|
|
|
|
if metaPartName == "service" && query.Service != "" {
|
|
return []byte(query.Service)
|
|
}
|
|
|
|
if metaPartName == "slo" && query.Slo != "" {
|
|
return []byte(query.Slo)
|
|
}
|
|
|
|
if metaPartName == "selector" && query.Selector != "" {
|
|
return []byte(query.Selector)
|
|
}
|
|
|
|
return in
|
|
})
|
|
|
|
return string(result)
|
|
}
|
|
|
|
func replaceWithMetricPart(metaPartName string, metricType string) []byte {
|
|
// https://cloud.google.com/monitoring/api/v3/metrics-details#label_names
|
|
shortMatches := metricNameFormat.FindStringSubmatch(metricType)
|
|
|
|
if metaPartName == "metric.name" {
|
|
if len(shortMatches) > 2 {
|
|
return []byte(shortMatches[3])
|
|
}
|
|
}
|
|
|
|
if metaPartName == "metric.service" {
|
|
if len(shortMatches) > 0 {
|
|
return []byte(shortMatches[1])
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func calcBucketBound(bucketOptions cloudMonitoringBucketOptions, n int) string {
|
|
bucketBound := "0"
|
|
if n == 0 {
|
|
return bucketBound
|
|
}
|
|
|
|
switch {
|
|
case bucketOptions.LinearBuckets != nil:
|
|
bucketBound = strconv.FormatInt(bucketOptions.LinearBuckets.Offset+(bucketOptions.LinearBuckets.Width*int64(n-1)), 10)
|
|
case bucketOptions.ExponentialBuckets != nil:
|
|
bucketBound = strconv.FormatInt(int64(bucketOptions.ExponentialBuckets.Scale*math.Pow(bucketOptions.ExponentialBuckets.GrowthFactor, float64(n-1))), 10)
|
|
case bucketOptions.ExplicitBuckets != nil:
|
|
bucketBound = fmt.Sprintf("%g", bucketOptions.ExplicitBuckets.Bounds[n])
|
|
}
|
|
return bucketBound
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) createRequest(ctx context.Context, dsInfo *models.DataSource, query *cloudMonitoringQuery, proxyPass string) (*http.Request, error) {
|
|
u, err := url.Parse(dsInfo.Url)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
u.Path = path.Join(u.Path, "render")
|
|
|
|
req, err := http.NewRequest(http.MethodGet, "https://monitoring.googleapis.com/", nil)
|
|
if err != nil {
|
|
slog.Error("Failed to create request", "error", err)
|
|
return nil, fmt.Errorf("failed to create request: %w", err)
|
|
}
|
|
|
|
req.Header.Set("Content-Type", "application/json")
|
|
req.Header.Set("User-Agent", fmt.Sprintf("Grafana/%s", setting.BuildVersion))
|
|
|
|
// find plugin
|
|
plugin, ok := plugins.DataSources[dsInfo.Type]
|
|
if !ok {
|
|
return nil, errors.New("unable to find datasource plugin CloudMonitoring")
|
|
}
|
|
|
|
var cloudMonitoringRoute *plugins.AppPluginRoute
|
|
for _, route := range plugin.Routes {
|
|
if route.Path == "cloudmonitoring" {
|
|
cloudMonitoringRoute = route
|
|
break
|
|
}
|
|
}
|
|
|
|
pluginproxy.ApplyRoute(ctx, req, proxyPass, cloudMonitoringRoute, dsInfo)
|
|
|
|
return req, nil
|
|
}
|
|
|
|
func (e *CloudMonitoringExecutor) getDefaultProject(ctx context.Context) (string, error) {
|
|
authenticationType := e.dsInfo.JsonData.Get("authenticationType").MustString(jwtAuthentication)
|
|
if authenticationType == gceAuthentication {
|
|
defaultCredentials, err := google.FindDefaultCredentials(ctx, "https://www.googleapis.com/auth/monitoring.read")
|
|
if err != nil {
|
|
return "", fmt.Errorf("failed to retrieve default project from GCE metadata server: %w", err)
|
|
}
|
|
token, err := defaultCredentials.TokenSource.Token()
|
|
if err != nil {
|
|
return "", fmt.Errorf("failed to retrieve GCP credential token: %w", err)
|
|
}
|
|
if !token.Valid() {
|
|
return "", errors.New("failed to validate GCP credentials")
|
|
}
|
|
|
|
return defaultCredentials.ProjectID, nil
|
|
}
|
|
return e.dsInfo.JsonData.Get("defaultProject").MustString(), nil
|
|
}
|