github.com/m3db/m3@v1.5.0/src/cmd/services/m3coordinator/downsample/flush_handler_test.go (about)

     1  // Copyright (c) 2018 Uber Technologies, Inc.
     2  //
     3  // Permission is hereby granted, free of charge, to any person obtaining a copy
     4  // of this software and associated documentation files (the "Software"), to deal
     5  // in the Software without restriction, including without limitation the rights
     6  // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
     7  // copies of the Software, and to permit persons to whom the Software is
     8  // furnished to do so, subject to the following conditions:
     9  //
    10  // The above copyright notice and this permission notice shall be included in
    11  // all copies or substantial portions of the Software.
    12  //
    13  // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    14  // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    15  // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    16  // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    17  // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
    18  // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
    19  // THE SOFTWARE.
    20  
    21  package downsample
    22  
    23  import (
    24  	"bytes"
    25  	"sync"
    26  	"testing"
    27  
    28  	"github.com/m3db/m3/src/metrics/metric/aggregated"
    29  	"github.com/m3db/m3/src/metrics/metric/id"
    30  	"github.com/m3db/m3/src/metrics/policy"
    31  	"github.com/m3db/m3/src/query/models"
    32  	"github.com/m3db/m3/src/query/storage/mock"
    33  	"github.com/m3db/m3/src/x/ident"
    34  	"github.com/m3db/m3/src/x/instrument"
    35  	"github.com/m3db/m3/src/x/pool"
    36  	"github.com/m3db/m3/src/x/serialize"
    37  	xsync "github.com/m3db/m3/src/x/sync"
    38  	xtest "github.com/m3db/m3/src/x/test"
    39  
    40  	"github.com/golang/mock/gomock"
    41  	"github.com/stretchr/testify/assert"
    42  	"github.com/stretchr/testify/require"
    43  	"github.com/uber-go/tally"
    44  )
    45  
    46  func TestDownsamplerFlushHandlerCopiesTags(t *testing.T) {
    47  	ctrl := xtest.NewController(t)
    48  	defer ctrl.Finish()
    49  
    50  	store := mock.NewMockStorage()
    51  	pool := serialize.NewMockMetricTagsIteratorPool(ctrl)
    52  
    53  	workers := xsync.NewWorkerPool(1)
    54  	workers.Init()
    55  
    56  	instrumentOpts := instrument.NewOptions()
    57  
    58  	handler := newDownsamplerFlushHandler(store, pool,
    59  		workers, models.NewTagOptions(), instrumentOpts)
    60  	writer, err := handler.NewWriter(tally.NoopScope)
    61  	require.NoError(t, err)
    62  
    63  	var (
    64  		expectedID = []byte("foo")
    65  		tagName    = []byte("name")
    66  		tagValue   = []byte("value")
    67  		annotation = []byte("annotation")
    68  	)
    69  	iter := serialize.NewMockMetricTagsIterator(ctrl)
    70  	gomock.InOrder(
    71  		iter.EXPECT().Reset(expectedID),
    72  		iter.EXPECT().NumTags().Return(1),
    73  		iter.EXPECT().Next().Return(true),
    74  		iter.EXPECT().Current().Return(tagName, tagValue),
    75  		iter.EXPECT().Next().Return(false),
    76  		iter.EXPECT().Err().Return(nil),
    77  		iter.EXPECT().Close(),
    78  	)
    79  
    80  	pool.EXPECT().Get().Return(iter)
    81  
    82  	// Write metric
    83  	err = writer.Write(aggregated.ChunkedMetricWithStoragePolicy{
    84  		ChunkedMetric: aggregated.ChunkedMetric{
    85  			ChunkedID:  id.ChunkedID{Data: expectedID},
    86  			TimeNanos:  123,
    87  			Value:      42.42,
    88  			Annotation: annotation,
    89  		},
    90  		StoragePolicy: policy.MustParseStoragePolicy("1s:1d"),
    91  	})
    92  	require.NoError(t, err)
    93  
    94  	// Wait for flush
    95  	err = writer.Flush()
    96  	require.NoError(t, err)
    97  
    98  	// Inspect the write
    99  	writes := store.Writes()
   100  	require.Equal(t, 1, len(writes))
   101  
   102  	// Ensure tag pointers _DO_NOT_ match but equal to same content
   103  	tags := writes[0].Tags().Tags
   104  	require.Equal(t, 1, len(tags))
   105  
   106  	tag := tags[0]
   107  	assert.True(t, bytes.Equal(tagName, tag.Name))
   108  	assert.True(t, bytes.Equal(tagValue, tag.Value))
   109  	assert.False(t, xtest.ByteSlicesBackedBySameData(tagName, tag.Name))
   110  	assert.False(t, xtest.ByteSlicesBackedBySameData(tagValue, tag.Value))
   111  
   112  	assert.Equal(t, annotation, writes[0].Annotation())
   113  }
   114  
   115  func graphiteTags(
   116  	t *testing.T, first string, encPool serialize.TagEncoderPool) []byte {
   117  	enc := encPool.Get()
   118  	defer enc.Finalize()
   119  
   120  	err := enc.Encode(ident.MustNewTagStringsIterator(
   121  		"__g0__", first,
   122  		"__g1__", "y",
   123  		"__g2__", "z",
   124  		string(MetricsOptionIDSchemeTagName), string(GraphiteIDSchemeTagValue),
   125  	))
   126  
   127  	require.NoError(t, err)
   128  	data, ok := enc.Data()
   129  	require.True(t, ok)
   130  	return append(make([]byte, 0, data.Len()), data.Bytes()...)
   131  }
   132  
   133  func TestDownsamplerFlushHandlerHighConcurrencyNoTagMixing(t *testing.T) {
   134  	ctrl := xtest.NewController(t)
   135  	defer ctrl.Finish()
   136  
   137  	store := mock.NewMockStorage()
   138  
   139  	size := 10
   140  	decodeOpts := serialize.NewTagDecoderOptions(serialize.TagDecoderOptionsConfig{
   141  		CheckBytesWrapperPoolSize: &size,
   142  	})
   143  
   144  	poolOpts := pool.NewObjectPoolOptions()
   145  	tagDecoderPool := serialize.NewTagDecoderPool(decodeOpts, poolOpts)
   146  	tagDecoderPool.Init()
   147  
   148  	pool := serialize.NewMetricTagsIteratorPool(tagDecoderPool, poolOpts)
   149  	pool.Init()
   150  
   151  	workers := xsync.NewWorkerPool(1)
   152  	workers.Init()
   153  
   154  	instrumentOpts := instrument.NewOptions()
   155  
   156  	handler := newDownsamplerFlushHandler(store, pool,
   157  		workers, models.NewTagOptions(), instrumentOpts)
   158  	writer, err := handler.NewWriter(tally.NoopScope)
   159  	require.NoError(t, err)
   160  
   161  	encodeOpts := serialize.NewTagEncoderOptions()
   162  	encPool := serialize.NewTagEncoderPool(encodeOpts, poolOpts)
   163  	encPool.Init()
   164  
   165  	xBytes := graphiteTags(t, "x", encPool)
   166  	fooBytes := graphiteTags(t, "foo", encPool)
   167  
   168  	var wg sync.WaitGroup
   169  	for i := 0; i < 100; i++ {
   170  		wg.Add(1)
   171  		xData := append(make([]byte, 0, len(xBytes)), xBytes...)
   172  		fooData := append(make([]byte, 0, len(fooBytes)), fooBytes...)
   173  		go func() {
   174  			defer wg.Done()
   175  			err := writer.Write(aggregated.ChunkedMetricWithStoragePolicy{
   176  				ChunkedMetric: aggregated.ChunkedMetric{
   177  					ChunkedID: id.ChunkedID{Data: xData},
   178  					TimeNanos: 123,
   179  					Value:     42.42,
   180  				},
   181  				StoragePolicy: policy.MustParseStoragePolicy("1s:1d"),
   182  			})
   183  			require.NoError(t, err)
   184  
   185  			err = writer.Write(aggregated.ChunkedMetricWithStoragePolicy{
   186  				ChunkedMetric: aggregated.ChunkedMetric{
   187  					ChunkedID: id.ChunkedID{Data: fooData},
   188  					TimeNanos: 123,
   189  					Value:     42.42,
   190  				},
   191  				StoragePolicy: policy.MustParseStoragePolicy("1s:1d"),
   192  			})
   193  			require.NoError(t, err)
   194  		}()
   195  	}
   196  
   197  	wg.Wait()
   198  	// Wait for flush
   199  	err = writer.Flush()
   200  	require.NoError(t, err)
   201  
   202  	// Inspect the write
   203  	writes := store.Writes()
   204  	require.Equal(t, 200, len(writes))
   205  
   206  	seenMap := make(map[string]int, 10)
   207  	for _, w := range writes {
   208  		str := w.Tags().String()
   209  		seenMap[str] = seenMap[str] + 1
   210  	}
   211  
   212  	assert.Equal(t, map[string]int{
   213  		"__g0__: foo, __g1__: y, __g2__: z": 100,
   214  		"__g0__: x, __g1__: y, __g2__: z":   100,
   215  	}, seenMap)
   216  }