github.com/m3db/m3@v1.5.0/src/cmd/services/m3comparator/main/querier.go (about)

     1  // Copyright (c) 2019 Uber Technologies, Inc.
     2  //
     3  // Permission is hereby granted, free of charge, to any person obtaining a copy
     4  // of this software and associated documentation files (the "Software"), to deal
     5  // in the Software without restriction, including without limitation the rights
     6  // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
     7  // copies of the Software, and to permit persons to whom the Software is
     8  // furnished to do so, subject to the following conditions:
     9  //
    10  // The above copyright notice and this permission notice shall be included in
    11  // all copies or substantial portions of the Software.
    12  //
    13  // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    14  // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    15  // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    16  // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    17  // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
    18  // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
    19  // THE SOFTWARE.
    20  
    21  package main
    22  
    23  import (
    24  	"bytes"
    25  	"context"
    26  	"fmt"
    27  	"math"
    28  	"math/rand"
    29  	"regexp"
    30  	"strconv"
    31  	"strings"
    32  	"sync"
    33  	"time"
    34  
    35  	"github.com/m3db/m3/src/cmd/services/m3comparator/main/parser"
    36  	"github.com/m3db/m3/src/dbnode/encoding"
    37  	"github.com/m3db/m3/src/dbnode/ts"
    38  	"github.com/m3db/m3/src/query/block"
    39  	"github.com/m3db/m3/src/query/models"
    40  	"github.com/m3db/m3/src/query/storage"
    41  	"github.com/m3db/m3/src/query/storage/m3"
    42  	"github.com/m3db/m3/src/query/storage/m3/consolidators"
    43  	xtime "github.com/m3db/m3/src/x/time"
    44  
    45  	"github.com/prometheus/common/model"
    46  )
    47  
    48  var _ m3.Querier = (*querier)(nil)
    49  
    50  type querier struct {
    51  	iteratorOpts         parser.Options
    52  	handler              seriesLoadHandler
    53  	blockSize            time.Duration
    54  	defaultResolution    time.Duration
    55  	histogramBucketCount uint
    56  	sync.Mutex
    57  }
    58  
    59  func newQuerier(
    60  	iteratorOpts parser.Options,
    61  	handler seriesLoadHandler,
    62  	blockSize time.Duration,
    63  	defaultResolution time.Duration,
    64  	histogramBucketCount uint,
    65  ) (*querier, error) {
    66  	if blockSize <= 0 {
    67  		return nil, fmt.Errorf("blockSize must be positive, got %d", blockSize)
    68  	}
    69  	if defaultResolution <= 0 {
    70  		return nil, fmt.Errorf("defaultResolution must be positive, got %d", defaultResolution)
    71  	}
    72  	return &querier{
    73  		iteratorOpts:         iteratorOpts,
    74  		handler:              handler,
    75  		blockSize:            blockSize,
    76  		defaultResolution:    defaultResolution,
    77  		histogramBucketCount: histogramBucketCount,
    78  	}, nil
    79  }
    80  
    81  func noop() error { return nil }
    82  
    83  func (q *querier) generateSeriesBlock(
    84  	start time.Time,
    85  	resolution time.Duration,
    86  	integerValues bool,
    87  ) parser.Data {
    88  	numPoints := int(q.blockSize / resolution)
    89  	dps := make(parser.Data, 0, numPoints)
    90  	for i := 0; i < numPoints; i++ {
    91  		stamp := start.Add(resolution * time.Duration(i))
    92  		var value float64
    93  		if integerValues {
    94  			value = float64(rand.Intn(1000))
    95  		} else {
    96  			value = rand.Float64()
    97  		}
    98  		dp := ts.Datapoint{
    99  			TimestampNanos: xtime.ToUnixNano(stamp),
   100  			Value:          value,
   101  		}
   102  
   103  		dps = append(dps, dp)
   104  	}
   105  
   106  	return dps
   107  }
   108  
   109  func (q *querier) generateSeries(
   110  	start time.Time,
   111  	end time.Time,
   112  	resolution time.Duration,
   113  	tags parser.Tags,
   114  	integerValues bool,
   115  ) (parser.IngestSeries, error) {
   116  	numBlocks := int(math.Ceil(float64(end.Sub(start)) / float64(q.blockSize)))
   117  	if numBlocks == 0 {
   118  		return parser.IngestSeries{}, fmt.Errorf("comparator querier: no blocks generated")
   119  	}
   120  
   121  	blocks := make([]parser.Data, 0, numBlocks)
   122  	for i := 0; i < numBlocks; i++ {
   123  		blocks = append(blocks, q.generateSeriesBlock(start, resolution, integerValues))
   124  		start = start.Add(q.blockSize)
   125  	}
   126  
   127  	return parser.IngestSeries{
   128  		Datapoints: blocks,
   129  		Tags:       tags,
   130  	}, nil
   131  }
   132  
   133  type seriesGen struct {
   134  	name string
   135  	res  time.Duration
   136  }
   137  
   138  // FetchCompressedResult fetches timeseries data based on a query.
   139  func (q *querier) FetchCompressedResult(
   140  	ctx context.Context,
   141  	query *storage.FetchQuery,
   142  	options *storage.FetchOptions,
   143  ) (consolidators.SeriesFetchResult, m3.Cleanup, error) {
   144  	var (
   145  		iters               encoding.SeriesIterators
   146  		randomSeries        []parser.IngestSeries
   147  		ignoreFilter        bool
   148  		err                 error
   149  		strictMetricsFilter bool
   150  	)
   151  
   152  	name := q.iteratorOpts.TagOptions.MetricName()
   153  	for _, matcher := range query.TagMatchers {
   154  		if bytes.Equal(name, matcher.Name) {
   155  
   156  			metricsName := string(matcher.Value)
   157  
   158  			// NB: the default behaviour of this querier is to return predefined metrics with random data if no match by
   159  			// metrics name is found. To force it return an empty result, query the "nonexistent*" metrics.
   160  			if match, _ := regexp.MatchString("^nonexist[ae]nt", metricsName); match {
   161  				return consolidators.SeriesFetchResult{}, noop, nil
   162  			}
   163  
   164  			if matcher.Type == models.MatchEqual {
   165  				strictMetricsFilter = true
   166  				iters, err = q.handler.getSeriesIterators(metricsName)
   167  				if err != nil {
   168  					return consolidators.SeriesFetchResult{}, noop, err
   169  				}
   170  
   171  				break
   172  			}
   173  		}
   174  	}
   175  
   176  	if iters == nil && !strictMetricsFilter && len(query.TagMatchers) > 0 {
   177  		iters, err = q.handler.getSeriesIterators("")
   178  		if err != nil {
   179  			return consolidators.SeriesFetchResult{}, noop, err
   180  		}
   181  	}
   182  
   183  	if iters == nil || iters.Len() == 0 {
   184  		randomSeries, ignoreFilter, err = q.generateRandomSeries(query)
   185  		if err != nil {
   186  			return consolidators.SeriesFetchResult{}, noop, err
   187  		}
   188  		iters, err = parser.BuildSeriesIterators(
   189  			randomSeries, xtime.ToUnixNano(query.Start), q.blockSize, q.iteratorOpts)
   190  		if err != nil {
   191  			return consolidators.SeriesFetchResult{}, noop, err
   192  		}
   193  	}
   194  
   195  	if !ignoreFilter {
   196  		filteredIters := filter(iters, query.TagMatchers)
   197  
   198  		cleanup := func() error {
   199  			iters.Close()
   200  			return nil
   201  		}
   202  
   203  		result, err := consolidators.NewSeriesFetchResult(
   204  			filteredIters, nil, block.NewResultMetadata())
   205  		return result, cleanup, err
   206  	}
   207  
   208  	cleanup := func() error {
   209  		iters.Close()
   210  		return nil
   211  	}
   212  
   213  	result, err := consolidators.NewSeriesFetchResult(
   214  		iters, nil, block.NewResultMetadata())
   215  	return result, cleanup, err
   216  }
   217  
   218  func (q *querier) generateRandomSeries(
   219  	query *storage.FetchQuery,
   220  ) (series []parser.IngestSeries, ignoreFilter bool, err error) {
   221  	var (
   222  		start = query.Start.Truncate(q.blockSize)
   223  		end   = query.End.Truncate(q.blockSize).Add(q.blockSize)
   224  	)
   225  
   226  	metricNameTag := q.iteratorOpts.TagOptions.MetricName()
   227  	for _, matcher := range query.TagMatchers {
   228  		if bytes.Equal(metricNameTag, matcher.Name) {
   229  			if matched, _ := regexp.Match(`^multi_\d+$`, matcher.Value); matched {
   230  				series, err = q.generateMultiSeriesMetrics(string(matcher.Value), start, end)
   231  				return
   232  			}
   233  			if matched, _ := regexp.Match(`^histogram_\d+_bucket$`, matcher.Value); matched {
   234  				series, err = q.generateHistogramMetrics(string(matcher.Value), start, end)
   235  				return
   236  			}
   237  		}
   238  	}
   239  
   240  	ignoreFilter = true
   241  	series, err = q.generateSingleSeriesMetrics(query, start, end)
   242  	return
   243  }
   244  
   245  func (q *querier) generateSingleSeriesMetrics(
   246  	query *storage.FetchQuery,
   247  	start time.Time,
   248  	end time.Time,
   249  ) ([]parser.IngestSeries, error) {
   250  	var (
   251  		gens = []seriesGen{
   252  			{"foo", time.Second},
   253  			{"bar", time.Second * 15},
   254  			{"quail", time.Minute},
   255  		}
   256  
   257  		actualGens []seriesGen
   258  	)
   259  
   260  	unlock := q.lockAndSeed(start)
   261  	defer unlock()
   262  
   263  	metricNameTag := q.iteratorOpts.TagOptions.MetricName()
   264  	for _, matcher := range query.TagMatchers {
   265  		// filter if name, otherwise return all.
   266  		if bytes.Equal(metricNameTag, matcher.Name) {
   267  			value := string(matcher.Value)
   268  			for _, gen := range gens {
   269  				if value == gen.name {
   270  					actualGens = append(actualGens, gen)
   271  					break
   272  				}
   273  			}
   274  
   275  			break
   276  		} else if "gen" == string(matcher.Name) {
   277  			cStr := string(matcher.Value)
   278  			count, err := strconv.Atoi(cStr)
   279  			if err != nil {
   280  				return nil, err
   281  			}
   282  
   283  			actualGens = make([]seriesGen, 0, count)
   284  			for i := 0; i < count; i++ {
   285  				actualGens = append(actualGens, seriesGen{
   286  					res:  q.defaultResolution,
   287  					name: fmt.Sprintf("foo_%d", i),
   288  				})
   289  			}
   290  
   291  			break
   292  		}
   293  	}
   294  
   295  	if len(actualGens) == 0 {
   296  		actualGens = gens
   297  	}
   298  
   299  	seriesList := make([]parser.IngestSeries, 0, len(actualGens))
   300  	for _, gen := range actualGens {
   301  		tags := parser.Tags{
   302  			parser.NewTag(model.MetricNameLabel, gen.name),
   303  			parser.NewTag("foobar", "qux"),
   304  			parser.NewTag("name", gen.name),
   305  		}
   306  
   307  		series, err := q.generateSeries(start, end, gen.res, tags, false)
   308  		if err != nil {
   309  			return nil, err
   310  		}
   311  
   312  		seriesList = append(seriesList, series)
   313  	}
   314  
   315  	return seriesList, nil
   316  }
   317  
   318  func (q *querier) generateMultiSeriesMetrics(
   319  	metricsName string,
   320  	start time.Time,
   321  	end time.Time,
   322  ) ([]parser.IngestSeries, error) {
   323  	suffix := strings.TrimPrefix(metricsName, "multi_")
   324  	seriesCount, err := strconv.Atoi(suffix)
   325  	if err != nil {
   326  		return nil, err
   327  	}
   328  
   329  	unlock := q.lockAndSeed(start)
   330  	defer unlock()
   331  
   332  	seriesList := make([]parser.IngestSeries, 0, seriesCount)
   333  	for id := 0; id < seriesCount; id++ {
   334  		tags := multiSeriesTags(metricsName, id)
   335  
   336  		series, err := q.generateSeries(start, end, q.defaultResolution, tags, false)
   337  		if err != nil {
   338  			return nil, err
   339  		}
   340  
   341  		seriesList = append(seriesList, series)
   342  	}
   343  
   344  	return seriesList, nil
   345  }
   346  
   347  func (q *querier) generateHistogramMetrics(
   348  	metricsName string,
   349  	start time.Time,
   350  	end time.Time,
   351  ) ([]parser.IngestSeries, error) {
   352  	suffix := strings.TrimPrefix(metricsName, "histogram_")
   353  	countStr := strings.TrimSuffix(suffix, "_bucket")
   354  	seriesCount, err := strconv.Atoi(countStr)
   355  	if err != nil {
   356  		return nil, err
   357  	}
   358  
   359  	unlock := q.lockAndSeed(start)
   360  	defer unlock()
   361  
   362  	seriesList := make([]parser.IngestSeries, 0, seriesCount)
   363  	for id := 0; id < seriesCount; id++ {
   364  		le := 1.0
   365  		var previousSeriesBlocks []parser.Data
   366  		for bucket := uint(0); bucket < q.histogramBucketCount; bucket++ {
   367  			tags := multiSeriesTags(metricsName, id)
   368  			leStr := "+Inf"
   369  			if bucket < q.histogramBucketCount-1 {
   370  				leStr = strconv.FormatFloat(le, 'f', -1, 64)
   371  			}
   372  			leTag := parser.NewTag("le", leStr)
   373  			tags = append(tags, leTag)
   374  			le *= 10
   375  
   376  			series, err := q.generateSeries(start, end, q.defaultResolution, tags, true)
   377  			if err != nil {
   378  				return nil, err
   379  			}
   380  
   381  			for i, prevBlock := range previousSeriesBlocks {
   382  				for j, prevValue := range prevBlock {
   383  					series.Datapoints[i][j].Value += prevValue.Value
   384  				}
   385  			}
   386  
   387  			seriesList = append(seriesList, series)
   388  
   389  			previousSeriesBlocks = series.Datapoints
   390  		}
   391  	}
   392  
   393  	return seriesList, nil
   394  }
   395  
   396  func multiSeriesTags(metricsName string, id int) parser.Tags {
   397  	return parser.Tags{
   398  		parser.NewTag(model.MetricNameLabel, metricsName),
   399  		parser.NewTag("id", strconv.Itoa(id)),
   400  		parser.NewTag("parity", strconv.Itoa(id%2)),
   401  		parser.NewTag("const", "x"),
   402  	}
   403  }
   404  
   405  func (q *querier) lockAndSeed(start time.Time) func() {
   406  	q.Lock()
   407  	rand.Seed(start.Unix())
   408  
   409  	return q.Unlock
   410  }
   411  
   412  // SearchCompressed fetches matching tags based on a query.
   413  func (q *querier) SearchCompressed(
   414  	ctx context.Context,
   415  	query *storage.FetchQuery,
   416  	options *storage.FetchOptions,
   417  ) (consolidators.TagResult, m3.Cleanup, error) {
   418  	return consolidators.TagResult{}, noop, fmt.Errorf("not impl")
   419  }
   420  
   421  // CompleteTagsCompressed returns autocompleted tag results.
   422  func (q *querier) CompleteTagsCompressed(
   423  	ctx context.Context,
   424  	query *storage.CompleteTagsQuery,
   425  	options *storage.FetchOptions,
   426  ) (*consolidators.CompleteTagsResult, error) {
   427  	nameOnly := query.CompleteNameOnly
   428  	// TODO: take from config.
   429  	return &consolidators.CompleteTagsResult{
   430  		CompleteNameOnly: nameOnly,
   431  		CompletedTags: []consolidators.CompletedTag{
   432  			{
   433  				Name:   []byte("__name__"),
   434  				Values: [][]byte{[]byte("foo"), []byte("foo"), []byte("quail")},
   435  			},
   436  		},
   437  	}, nil
   438  }